Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

LlaMa was trained on 78 GB of StackExchange (I assume StackOverflow was included in that).


But was it parsed and reformatted specifically in the "chat format" (i.e. the same as inputs later fed to the model when used as a chatbot)? It can make a surprisingly big difference.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: