top | item 47088066

(no title)

enceladus06 | 9 days ago

OpenAI and Anthropic don't have a moat. We will have actual open models like DeepSeek and Kimi with the same functionality as Opus 4.6 in Claude Code <6mo IMO. Competition is a good thing for the end-user.

discuss

order

zozbot234|9 days ago

The open-weight models are great but they're roughly a full year behind frontier models. That's a lot. There's also a whole lot of uses where running a generic Chinese-made model may be less than advisable, and OpenAI/Anthropic have know-how for creating custom models where appropriate. That can be quite valuable.

coder543|9 days ago

I would not say a full year... not even close to a year: GLM-5 is very close to the frontier: https://artificialanalysis.ai/

Artificial Analysis isn't perfect, but it is an independent third party that actually runs the benchmarks themselves, and they use a wide range of benchmarks. It is a better automated litmus test than any other that I've been able to find in years of watching the development of LLMs.

And the gap has been rapidly shrinking: https://www.youtube.com/watch?v=0NBILspM4c4&t=642s

mattmaroon|9 days ago

That's a lot now, in the same way that a PC in 1999 vs a PC in 2000 was a fairly sizeable discrepancy. At some point, probably soon, progress will slow, and it won't be much.

jnovek|9 days ago

I just did a test project using K2.5 on opencode and, for me, it doesn’t even come close to Claude Code. I was constantly having to wrangle the model to prevent it from spewing out 1000 lines at once and it couldn’t hold the architecture in its head so it would start doing things in inconsistent ways in different parts of the project. What it created would be a real maintenance nightmare.

It’s much better than the previous open models but it’s not yet close.