top | item 46905055

(no title)

boroboro4 | 24 days ago

While I mostly agree with you, it worth noting modern llms are trained on 10-20-30T of tokens which is quite comparable to their size (especially given how compressible the data is)

discuss

order

No comments yet.