top | item 46905055 (no title) boroboro4 | 24 days ago While I mostly agree with you, it worth noting modern llms are trained on 10-20-30T of tokens which is quite comparable to their size (especially given how compressible the data is) discuss order hn newest No comments yet.
No comments yet.