Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Enjoy the insight, but the title makes my eye twitch. How about "LLM weights are pieces of history"?


Small LLM weights are not really interesting though. I am currently training GPT-2 small sized models for a scientific project right, and their world models are just not good enough to generate any kind of real insight about the world it was trained in except for corpus biases.


Small large language models? This sounds like the apocryphal headline when a spiritualist with dwarfism escaped prison: "Small medium at large." Do you also have some dehydrated water and a secure key escrow system?


A collection of newspapers is generally a better source than a single leaflet, but even a leaflet is a piece of history.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: