(no title)
intrasight | 3 days ago
For models that run on general-purpose AI hardware, I don't know why the vendors would waste that resource on old models.
intrasight | 3 days ago
For models that run on general-purpose AI hardware, I don't know why the vendors would waste that resource on old models.
vidarh|3 days ago
In terms of price, I can get 1m output tokens from Deepseek for 40 cents vs. 25 dollars for Opus, and a number of models near the 1-2 dollar mark that are increasingly viable for a larger set of applications.
Providers will keep running those cheaper models as long as there's demand.
generallyjosh|2 days ago
And, depending on effort settings, they do more 'thinking', i.e., use more rounds of inference to generate longer internal chains of thought
Both very good reasons to prefer a smaller model, if the small model is good enough for the task