top | item 44822263

(no title)

Kurtz79 | 6 months ago

Does it even make sense calling them 'GPUs' (I just checked NVIDIA product page for the H100 and it is indeed so)?

There should be a quicker way to differentiate between 'consumer-grade hardware that is mainly meant to be used for gaming and can also run LLMs inference in a limited way' and 'business-grade hardware whose main purpose is AI training or running inference for LLMs".

discuss

order

blitzar|6 months ago

We are fast approaching the return of the math coprocessor. In fashion they say that trends tend to reappear roughly every two decades, its overdue.

egorfine|6 months ago

Yeah I would love for Nvidia to introduce faster update cycle to their hardware, so that we'll have models like "H201", "H220", etc.

I think it will also make sense to replace "H" with a brand number, sort of like they already do for customer GPUs.

So then maybe one day we'll have a math coprocessor called "Nvidia 80287".

beAbU|6 months ago

I remember the building hugh end workstations for a summer job in the 2000s, where I had to fit Tesla cards in the machines. I don't remember what their device names were, we just called them tesla cards.

"Accelerator card" makes a lot of sense to me.

WithinReason|6 months ago

It's called a tensorcore and it's in most GPUs

genewitch|6 months ago

"GPGPU" was something from over a decade ago; for general purpose GPU computing

hnuser123456|6 months ago

Yeah, Crysis came out in 2007 and could run physics on the GPU.

AlphaSite|6 months ago

I think apple calls them NPUs and Broadcom calls them XPUs. Given they’re basically the number 2 and 3 accelerator manufacturers one of those probably works.

codedokode|6 months ago

By the way I wonder, what has more performance, a $25 000 professional GPU or a bunch of cheaper consumer GPUs costing $25 000 in total?

omneity|6 months ago

Consumer GPUs in theory and by a large margin (10 5090s will eat an H100 lunch with 6 times the bandwidth, 3x VRAM and a relatively similar compute ratio), but your bottleneck is the interconnect and that is intentionally crippled to avoid beowulf GPU clusters eating into their datacenter market.

Last consumer GPU with NVLink was the RTX 3090. Even the workstation-grade GPUs lost it.

https://forums.developer.nvidia.com/t/rtx-a6000-ada-no-more-...

washadjeffmad|6 months ago

I just specify SXM (node) when I want to differentiate from PCIe. We have H100s in both.

addandsubtract|6 months ago

We could call the consumer ones GFX cards, and keep GPU for the matrix multiplying ones.

beAbU|6 months ago

GPU stands for "graphics processing unit" so I'm not sure how your suggestion solves it.

Maybe renaming the device to an MPU, where the M stands for "matrix/math/mips" would make it more semantically correct?

amelius|6 months ago

Well, does it come with graphics connectors?

OliverGuy|6 months ago

Nope, doesn't have any of the required hardware to even process graphics iirc