This is true, but only if you have a GPU (/accelerator) comparable in performance to the one backing the service, or at least comparable after accounting for the local benefit. This is an expensive proposition because it will be sitting idle between completions and when you're not coding.
daemonologist|1 year ago
sqs|1 year ago
I just used Cody with Ollama for local inference on a flight where the wifi was broken, and it never fails to blow my mind: https://x.com/sqs/status/1803269013310759236.
rdedev|1 year ago
ado__dev|1 year ago
https://sourcegraph.com/blog/local-code-completion-with-olla...
s1mplicissimus|1 year ago
morgante|1 year ago