top | item 46903432

(no title)

ponyous | 24 days ago

I think models are smart enough for most of the stuff, these little incremental changes barely matter now. What I want is the model that is fast.

discuss

order

energy123|24 days ago

I predict a bifurcation in usage.

Serial usecases ("fix this syntax errors") will go on Cerebras and get 10x faster.

Deep usecases ("solve Riemann hypothesis") will become massively parallel and go on slower inference compute.

Teams will stitch both together because some workflows go through stages of requiring deep parallel compute ("scan my codebase for bugs and propose fixes") followed by serial compute ("dedupe and apply the 3 fixes, resolve merge conflict").

newtwilly|24 days ago

I've been using 5.1-codex-max with low reasoning (in Cursor fwiw) recently and it feels like a nice speed while still being effective. Might be worth a shot.

derac|24 days ago

This is faster if their marketing is right, it uses significantly less tokens. Gemini 3 flash is very good as well.