As far as I understand all the inference purpose-build silicon out there is not being sold to competitors and kept in-house. Google's TPU, Amazon's Inferentia (horrible name), Microsoft's Maia, Meta's MTIA. It seems that custom inference silicon is a huge part of the AI game. I doubt GPU-based inference will be relevant/competitive soon.
nightshift1|1 month ago
https://newsletter.semianalysis.com/p/tpuv7-google-takes-a-s...
nomel|1 month ago
Is there any public info about % inference on custom vs GPU, for these companies?
mrinterweb|1 month ago
almostgotcaught|1 month ago
mrinterweb|1 month ago