>So other than training ever-larger models on the same internet data, how can they make better LLMs?
Training a multi-modal model that can integrate audio, visual, text and all sorts of data modalities to human level capabilities still remains an clear challenge. The bottleneck here is not the lack of data imo.
No comments yet.