(no title)
danrocks | 2 years ago
I don't think this is accurate. The Stanford team used LLaMA as base model and added a smaller model on top of it - training the joint model using data (generated from ChatGPT) is what cost $600. Nobody trained a GPT-like model from scratch for $600 - this experiment took advantage of the millions of USD used to train the larger models.
No comments yet.