Hacker Read
top
|
best
|
new
|
newcomments
|
leaders
|
about
|
bookmarklet
login
GaggiX | karma 3608 | avg karma 2.43
2023-04-17 09:56:47
|
next
[–]
update item
All Pythia models were trained on 300B tokens, LLaMa models were trained on 1/1.4T tokens.
reply
view as:
tree
latest_first
Legal
|
privacy
reply