Show HN: First Claude Code client for Ollama local models
44 points| SerafimKorablev | 1 month ago |github.com
Here is the release note from Ollama that made this possible: https://ollama.com/blog/claude
Technically, what I do is pretty straightforward:
- Detect which local models are available in Ollama.
- When internet access is unavailable, the client automatically switches to Ollama-backed local models instead of remote ones.
- From the user’s perspective, it is the same Claude Code flow, just backed by local inference.
In practice, the best-performing model so far has been qwen3-coder:30b. I also tested glm-4.7-flash, which was released very recently, but it struggles with reliably following tool-calling instructions, so it is not usable for this workflow yet.
oceanplexian|1 month ago
davely|1 month ago
xd1936|1 month ago
d4rkp4ttern|1 month ago
https://github.com/pchalasani/claude-code-tools/blob/main/do...
One tricky thing that took me a whole day to figure out is that using Claude Code in this setup was causing total network failures due to telemetry pings, so I had to set this env var to 1: CLAUDE_CODE_DISABLE_NONESSENTIAL_TRAFFIC
eli|1 month ago
g4cg54g54|1 month ago
in particular i´d like to call claude-models - in openai-schema hosted by a reseller - with some proxy that offers anthropic format to my claude --- but it seems like nothing gets to fully line things up (double-translated tool names for example)
reseller is abacus.ai - tried BerriAI/litellm, musistudio/claude-code-router, ziozzang/claude2openai-proxy, 1rgs/claude-code-proxy, fuergaosi233/claude-code-proxy,
dsrtslnd23|1 month ago
derp-mcgee|1 month ago
thtmnisamnstr|1 month ago
ryandrake|1 month ago
horacemorace|1 month ago
g4cg54g54|1 month ago
d4rkp4ttern|1 month ago
But with Qwen3-30B-A3B I get 20 tps in CC.
dosinga|1 month ago
d0100|1 month ago
smissingham|1 month ago
[deleted]
mchiang|1 month ago
https://github.com/21st-dev/1code
dang|1 month ago