NVIDIA's H100 Hopper GPU Sweeps MLPerf AI Inference Tests With Huge Performance Uplift
NVIDIA is taking generative AI and other workloads to new heights with its H100 and L4 GPUs. The latest MLPerf 3.0 test results highlight Hopper delivering 4x more performance than A100.
The 2023 GTC Developer conference included 650 talks from leaders in AI development. One of the most prominent of those was from NVIDIA's own Jensen Huang. Huang's keynote speech covered how the company and its partners are offering everything from training to deployment for cutting-edge services. Included in all that is NVIDIA's new H100 NVL with dual-GPU NVLink. The new GPU is based on NVIDIA's Hopper architecture and features a Transformer Engine design.
"The next level of Generative AI requires new AI infrastructure to train large language models with great energy efficiency. Customers are ramping Hopper at scale, building AI infrastructure with tens of thousands of Hopper GPUs connected by NVIDIA NVLink and InfinBand," Huang remarked in a recent NVIDIA blog post.
L4 Tensor Core GPUs are also seeing a 3x in performance increase over last-generation hardware. NVIDIA pointed out that the L4 was particularly impressive on the performance-hungry BERT model, due to it supporting the key FP8 format.
Ten companies submitted results on the NVIDIA platform during this round of MLPerf testing. Those included Microsoft Azure cloud service and system makers including ASUS, Dell Technologies, Gigabyte, H3C, Lenovo, Nettrix, Supermicro and xFusion.
Anyone wanting to take a deeper dive into the MLPerf results can do so by visiting NVIDIA's developer website.