top | item 46235887

(no title)

cc62cf4a4f20 | 2 months ago

In other news, been using Devstral 2 (Ollama) with OpenCode, and while it's not as good as Claude Code, my initial sense it that it's nonetheless good enough and doesn't require me to send my data off my laptop.

I kind of wonder how close we are to alternative (not from a major AI lab) models being good enough for a lot of productive work and data sovereignty being the deciding factor.

discuss

order

Nesco|2 months ago

Wait, isn't Devstral2 (normal not small) 123b? What type of laptop do you have? MacBooks don't go over 128GiB

cc62cf4a4f20|2 months ago

I'm using small - works well for its size

yberreby|2 months ago

Would you share some additional details? CPU, amount of unified memory / VRAM? Tok/s with those?

cc62cf4a4f20|2 months ago

MBP M4 Max 64MB - haven't measured the tokens/sec, feels slower than Claude, but not unbearably

It's not yet perfect, my sense is just that it's near the tipping point where models are efficient enough that running a local model is truly viable