(no title)
vidarh | 3 days ago
For some tasks that matters. But for a lot of tasks, "good enough but cheaper" will win out.
I'm sure there will be a market for whichever company has the best model, but just like most companies don't hire many PhD's, most companies won't feel a need for the highest end models either, above a certain level.
E.g. with the release of Sonnet 4.6, I switched a lot of my processes from Opus to Sonnet, because Sonnet 4.6 is good enough, and it means I can do more for less.
But I'm also experimenting with Kimi, Qwen, Deepseek, and others for a number of tasks, including fine-grained switching and interleaving. E.g. have a cheap but dumb model filter data or take over when a sub-task is simple enough, in order to have the smart model do less, for example.
intrasight|3 days ago
For models that run on general-purpose AI hardware, I don't know why the vendors would waste that resource on old models.
vidarh|3 days ago
In terms of price, I can get 1m output tokens from Deepseek for 40 cents vs. 25 dollars for Opus, and a number of models near the 1-2 dollar mark that are increasingly viable for a larger set of applications.
Providers will keep running those cheaper models as long as there's demand.
generallyjosh|2 days ago
And, depending on effort settings, they do more 'thinking', i.e., use more rounds of inference to generate longer internal chains of thought
Both very good reasons to prefer a smaller model, if the small model is good enough for the task