Hacker Read top | best | new | newcomments | leaders | about | bookmarklet login

75B tokens is not really enough data to make an intelligent model. Llama was trained on over 1 trillion tokens.

And yes, training on top of LLaMA could introduce a lot of unexpected behavior, but that's just where the State-of-the-Art is today



view as:

Legal | privacy