Ask HN: 2x Arc A770 or 1x Radeon 7900 XTX for llama.cpp
5 points| danielEM | 11 months ago
From what I dig so far it looks like dual Arc A770 is supported by llama.cpp. And saw some reports that llama.cpp on top of IPEX-LLM is fastest way for inference on intel card.
On the other end there is more expensive 7900 XTX on which AMD claims (Jan '25) that inference is faster than on 4090.
So - what is the state of the art as of today, how does one compare to another (apple to apple)? What is tokens/s diff?
runjake|11 months ago
https://www.reddit.com/r/LocalLLaMA/
Using the search gave me a bunch of threads, but here's one:
https://www.reddit.com/r/LocalLLaMA/comments/1ip6c9e/looking...
danielEM|11 months ago
laweijfmvo|11 months ago
danielEM|11 months ago
unknown|11 months ago
[deleted]