How? I have an M2 Pro and I run 7B and 13B models through Ollama and also LM Studio.
Because there’s no CUDA, the speed is much slower than ChatGPT. The answers from 7B are also not at the same quality as ChatGPT. (Lots of mistakes and hallucinations)
wenc|2 years ago
Because there’s no CUDA, the speed is much slower than ChatGPT. The answers from 7B are also not at the same quality as ChatGPT. (Lots of mistakes and hallucinations)
bdavbdav|2 years ago