Exactly the same has been said over and over again, ever since CUDA took off for scientific computing around 2010. I don’t really understand why 15 years later AMD still hasn’t been able to copy the recipy, and frankly it may be too late now with all that mindshare in NVIDIA’s software stack.
bayindirh|8 months ago
NVIDIA has a moat for smaller systems, but that is not true for clusters.
As long as you have a team to work with the hardware you have, performance beats mindshare.
aseipp|8 months ago
Nvidia of course has a shitload more money, and they've been doing this for longer, but that's just life.
> smaller systems
El Capitan is estimated to cost around $700 million or something with like 50k deployed MI300 GPUs. xAI's Colossus cluster alone is estimated to be north of $2 billion with over 100k GPUs, and that's one of ~dozens of deployed clusters Nvidia has developed in the past 5 years. AI is a vastly bigger market in every dimension, from profits to deployments.
wmf|8 months ago
pjmlp|8 months ago
bigyabai|8 months ago
7speter|8 months ago