(no title)
Muskyinhere | 1 year ago
But obvoiusly they don't.
And for reasons: NVidia has worked on CUDA for ages, do you believe they just replace this whole thing in no time?
Muskyinhere | 1 year ago
But obvoiusly they don't.
And for reasons: NVidia has worked on CUDA for ages, do you believe they just replace this whole thing in no time?
Wytwwww|1 year ago
Der_Einzige|1 year ago
Or you can just buy Nvidia.
treprinum|1 year ago
m00x|1 year ago
llama.cpp is just inference, not training, and the CUDA backend is still the fastest one by far. No one is even close to matching CUDA on either training or inference. The closest is AMD with ROCm, but there's likely a decade of work to be done to be competitive.
pjmlp|1 year ago