top | item 35294040

(no title)

michaelhartm | 2 years ago

They used the 6b GPT4-J, not 20B. That's what's interesting, it's a smallish large language model :).

discuss

order