(no title)
lm28469 | 1 day ago
You're comparing 100b parameters open models running on a consumer laptop VS private models with at the very least 1t parameters running on racks of bleeding edge professional gpus
Local agentic coding is closer to "shit me the boiler plate for an android app" not "deep research questions", especially on your machine
vlovich123|1 day ago
Speculation is that the frontier models are all below 200B parameters but a 2x size difference wouldn’t fully explain task performance differences
nl|22 hours ago
Some versions of some the models are around that size, which you might hit for example with the ChatGPT auto-router.
But the frontier models are all over 1T parameters. Source: watch interview with people who have left one of the big three labs and now work at the Chinese labs and are talking about how to train 1T+ models.
BoredomIsFun|15 hours ago
NamlchakKhandro|23 hours ago
Yes it does.
827a|22 hours ago
Core speed/count and memory bandwidth determines your performance. Memory size determines your model size which determines your smarts. Broadly speaking.
ses1984|1 day ago
shlomo_z|19 hours ago
mstaoru|16 hours ago
delaminator|1 day ago
lm28469|1 day ago
There are the benchmarks, the promises, and what everybody can try at home