Microsoft Shares Inside Look At NVIDIA's Supercomputing AI Tech Powering ChatGPT
"Co-designing supercomputers with Azure has been crucial for scaling our demanding AI training needs, making our research and alignment work on systems like ChatGPT possible," stated Greg Brockman, President and Co-Founder of OpenAI.
The ND H100 v5 Virtual Machine (VM) enables on-demand scaling in sizes ranging from eight to thousands of NVIDIA H100 GPUs interconnected by NVIDIA Quantum-2 InfiniBand networking. Microsoft says this will equate to customers seeing "significantly faster performance" for AI models over the last generation.
“NVIDIA and Microsoft Azure have collaborated through multiple generations of products to bring leading AI innovations to enterprises around the world. The NDv5 H100 virtual machines will help power a new era of generative AI applications and services," remarked Ian Buck, Vice President of hyperscale and high-performance computing at NVIDIA.
Chappell added that there is a lot of system-level optimization needed in order to get the best performance. That optimization includes software that facilitates effective utilization of the GPUs and networking equipment. The Azure infrastructure is currently optimized for large language model training and is available through Azure AI supercomputing capabilities in the cloud.
“This is the most extraordinary moment we have witnessed in the history of AI,” Jensen Huang remarked in a recent NVIDIA blog post. “New AI technologies and rapidly spreading adoption are transforming science and industry, and opening new frontiers for thousands of new companies. This will be our most important GTC yet.”
Microsoft touts that only Microsoft Azure offers the GPUs, the InfiniBand networking, and the unique AI infrastructure needed to build such transformational AI models at scale.