(no title)
scriptsmith | 2 years ago
I see the GitHub copilot extensions gets a new release one every few days, so is it just that the way they're integrated is more complicated so not worth the effort?
scriptsmith | 2 years ago
I see the GitHub copilot extensions gets a new release one every few days, so is it just that the way they're integrated is more complicated so not worth the effort?
sestinj|2 years ago
thewataccount|2 years ago
This works well for me except the 15B+ don't run fast enough on a 4090 - hopefully exllama supports non-llama models, or maybe it'll support CodeLLaMa already I'm not sure.
For general chat testing/usage this works pretty well with lots of options - https://github.com/oobabooga/text-generation-webui/
msp26|2 years ago
I assume quantized models will run a lot better. TheBloke already seems like he's on it.
https://huggingface.co/TheBloke/CodeLlama-13B-fp16
modeless|2 years ago
lhl|2 years ago
fudged71|2 years ago