Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

GPT-J was the one that made me really interested in LLMs, as I could run it on a 3090.

Some details on the timeline are not quite precise, and would benefit from linking to a source so that everyone can verify it. For example, HyperClOVA is listed as 204B parameters, but it seems it used 560B parameters (https://aclanthology.org/2021.emnlp-main.274/).

 help



Great idea! Thanks



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: