(no title)
htsh | 1 year ago
ollama run deepseek-r1:14b
generally, if the model file size < your vram, it is gonna run well. this file is 9gb.
if you don't mind slower generation, you can run models that fit within your vram + ram, and ollama will handle that offloading of layers for you.
so the 32b should run on your system, but it is gonna be much slower as it will be using GPU + CPU.
prob of interest: https://simonwillison.net/2025/Jan/20/deepseek-r1/
-h
jordiburgos|1 year ago
I am testing it now and seems quite fast giving the responses for a local model.