Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

LLAMA is an open-weights model. I like this term, let's use that instead of open source.


Can a human programmer edit the weights according to some semantics?


It is possible to merge two fine-tunes of models from the same family by... wait for it... averaging or combining their weights[0].

I am still amazed that we can do that.

[0]: https://arxiv.org/abs/2212.09849


This is absolutely wild.


Yes. Using fine tuning.


Yes, there is the concept of a "frakenmerge" and folks have also bolted on vision and audio models to LLMs.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: