NVIDIA Announces Scalable GPU-Accelerated Supercomputer in the Microsoft Azure Cloud
New Microsoft Azure NDv2 supersized instance can scale to hundreds of interconnected NVIDIA Tensor Core GPUs for complex AI and high-performance computing applications.
November 21, 2019
Built to handle demanding artificial intelligence (AI) and high-performance computing applications, the largest deployments of Azure’s new NDv2 instance rank among the world’s fastest supercomputers, offering up to 800 NVIDIA V100 Tensor Core GPUs interconnected on a single Mellanox InfiniBand backend network, NVIDIA reports. It enables customers for the first time to rent an entire AI supercomputer on demand from their desk, and match the capabilities of large-scale, on-premises supercomputers that can take months to deploy.
“Until now, access to supercomputers for AI and high-performance computing has been reserved for the world’s largest businesses and organizations,” says Ian Buck, vice president and general manager of Accelerated Computing at NVIDIA. “Microsoft Azure’s new offering democratizes AI, giving wide access to an essential tool needed to solve some of the world’s biggest challenges.”
“As cloud computing gains momentum everywhere, customers are seeking more powerful services,” adds Girish Bablani, corporate vice president of Azure Compute at Microsoft Corp. “Working with NVIDIA, Microsoft is giving customers instant access to a level of supercomputing power that was previously unimaginable, enabling a new era of innovation.”
The new offering—which is ideal for complex AI, machine learning and HPC workloads—can provide performance and cost advantages over traditional CPU-based computing. AI researchers needing fast solutions can quickly spin up multiple NDv2 instances and train complex conversational AI models in just hours.
Microsoft and NVIDIA engineers used 64 NDv2 instances on a pre-release version of the cluster to train BERT, a conversational AI model, in roughly three hours. This was achieved in part by taking advantage of multi-GPU optimizations provided by NCCL, an NVIDIA CUDA X library and high-speed Mellanox interconnects.
Customers can also see benefits from using multiple NDv2 instances to run complex HPC workloads, such as LAMMPS, a molecular dynamics application used to simulate materials down to the atomic scale in such areas as drug development and discovery. A single NDv2 instance can deliver an order of magnitude faster results than a traditional HPC node without GPU acceleration for specific types of applications, such as deep learning. This performance can scale linearly to a hundred instances for large-scale simulations.
All NDv2 instances benefit from the GPU-optimized HPC applications, machine learning software and deep learning frameworks like TensorFlow, PyTorch and MXNet from the NVIDIA NGC container registry and Azure Marketplace. The registry also offers Helm charts to easily deploy the AI software on Kubernetes clusters.
Availability and Pricing
NDv2 is available now in preview. One instance with eight NVIDIA V100 GPUs can be clustered to scale up to a variety of workload demands.
Sources: Press materials received from the company and additional information gleaned from the company’s website.
More NVIDIA Coverage
About the Author
DE’s editors contribute news and new product announcements to Digital Engineering.
Press releases may be sent to them via DE-Editors@digitaleng.news.