Settings

Theme

Ask HN: Are Nvidia H100s that good?

4 points by jonathanlei 2 years ago · 5 comments · 1 min read

Reader

I was playing around with some different GPUs yesterday and put all of the results here: https://www.tensordock.com/benchmarks

I tried a vLLM and Resnet training workload. The H100 outperforms the A100 about 45% to 80% consistently, but it isn’t that much faster…

What workloads would see the most speedup, because I’m really not seeing 3x+ any on vLLM or simple training workloads?

zer00eyz 2 years ago

You're doing inference, not training.

https://lambdalabs.com/gpu-benchmarks

  • utopcell 2 years ago

    Interesting summarization. For throughput/watt, nothing beats A100 40GB PCIe cards. In terms of throughput/$, 4090 cards are >8X better than the best H100.

  • jonathanleiOP 2 years ago

    Hmm I did include a training workload as the second chart. My test workload was relatively small so I guess if the workload I ran spends a bit less GPU time comparatively to the CPU, given equal CPU for all workloads, would be an equalizing factor.

    But even looking at the Lambda Labs benchmarks, I am surprised that the H100 PCIE barely outperforms the A100 SXM, for example. And it is meant to be a replacement for the A100 PCIE. 20% generational improvement yes, but I would have expected more?

    • zer00eyz 2 years ago

      >> My test workload was relatively small

      This is the game changer. More memory and more interconnect speed = better

      >> H100 PCIE barely outperforms the A100 SXM

      This is the better interconnect... its only useful if your using it. IF you can fit your workload in the 80gb of the H100 then the SXM becomes far less useful.

jonathanleiOP 2 years ago

Oops, just noticed the link isn’t clickable, here you go! https://tensordock.com/benchmarks

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection