top | item 40721762

(no title)

shahahmed | 1 year ago

arguably you can reduce even more latency by keeping the model on-device as well, but that would mean revealing the weights of the fine-tuned model.

If the user preferred reduced latency and had the RAM, is that an option?

discuss

order

daemonologist|1 year ago

This is true, but only if you have a GPU (/accelerator) comparable in performance to the one backing the service, or at least comparable after accounting for the local benefit. This is an expensive proposition because it will be sitting idle between completions and when you're not coding.

s1mplicissimus|1 year ago

the model is probably most of the "secret sauce" of cody, so if they gave that away people could just copy it around like mp3s. my guess

morgante|1 year ago

Completely incorrect, as Sourcegraph has not historically trained models and Cody swaps between many open source and 3rd party models.