Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

"Second, given the large gap between LLaMA and ChatGPT (the latter model is faster, cheaper, and more accurate), "

No it's not, llama would be cheaper and likely faster if you ran it on the same scale, actually there've been a few calcs done, that running llama 65b if you're at 100% usage is cheaper than 3.5turbo per token. Also comparing them for accuracy isn't fair comparison, one is a foundational model, one is an instruct tuned model. Perhaps compare llama 65b with gpt3.



Isn't it the comparison is like "My home PC server is far cheaper than EC2"?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: