(no title)
7734128 | 24 days ago
This is almost like asking me to invent a path finding algorithm when I've been thought Dijkstra's and A*.
7734128 | 24 days ago
This is almost like asking me to invent a path finding algorithm when I've been thought Dijkstra's and A*.
NitpickLawyer|24 days ago
A pertinent quote from the article (which is a really nice read, I'd recommend reading it fully at least once):
> Previous Opus 4 models were barely capable of producing a functional compiler. Opus 4.5 was the first to cross a threshold that allowed it to produce a functional compiler which could pass large test suites, but it was still incapable of compiling any real large projects. My goal with Opus 4.6 was to again test the limits.
wmf|24 days ago
simonw|24 days ago
lossolo|24 days ago
How many agents did they use with previous Opus? 3?
You've chosen an argument that works against you, because they actually could do that if they were trained to.
Give them the same post-training (recipes/steering) and the same datasets, and voila, they'll be capable of the same thing. What do you think is happening there? Did Anthropic inject magic ponies?
f311a|23 days ago
And keep in mind, the original creators of the first compiler had to come up with everything: lexical analysis -> parsing -> IR -> codegen -> optimization. LLMs are not yet capable of producing a lot of novelty. There are many areas in compilers that can be optimized right now, but LLMs can't help with that.
fatherwavelet|24 days ago
Then they start improvising and the same person counters with "what a bunch of slop, just making things up!"
falloutx|24 days ago
calebhwin|24 days ago
[deleted]
zephen|24 days ago
They only have to keep reiterating this because people are still pretending the training data doesn't contain all the information that it does.
> It's not like any LLM could 1for1 regurgitate millions of LoC from any training set... This is not how it works.
Maybe not any old LLM, but Claude gets really close.
https://arxiv.org/pdf/2601.02671v1
skydhash|24 days ago
lunar_mycroft|24 days ago
(I'm not claiming this is what actually happened here, just pointing out that memorization is a lot more plausible/significant than you say)
[0] https://www.theregister.com/2026/01/09/boffins_probe_commerc...