GTC—NVIDIA today announced NVIDIA DGX™ Cloud, an AI supercomputing service that gives enterprises immediate access to the infrastructure and software needed to train advanced models for generative AI and other groundbreaking applications.
DGX Cloud provides dedicated clusters of NVIDIA DGX AI supercomputing, paired with NVIDIA AI software. The service makes it possible for every enterprise to access its own AI supercomputer using a simple web browser, removing the complexity of acquiring, deploying and managing on-premises infrastructure.
Enterprises rent DGX Cloud clusters on a monthly basis, which ensures they can quickly and easily scale the development of large, multi-node training workloads without having to wait for accelerated computing resources that are often in high demand.
“We are at the iPhone moment of AI. Startups are racing to build disruptive products and business models, and incumbents are looking to respond,” said Jensen Huang, founder and CEO of NVIDIA. “DGX Cloud gives customers instant access to NVIDIA AI supercomputing in global-scale clouds.”
NVIDIA is partnering with leading cloud service providers to host DGX Cloud infrastructure, starting with Oracle Cloud Infrastructure (OCI). Its OCI Supercluster provides a purpose-built RDMA network, bare-metal compute and high-performance local and block storage that can scale to superclusters of over 32,000 GPUs.
Microsoft Azure is expected to begin hosting DGX Cloud next quarter, and the service will soon expand to Google Cloud and more.
Industry Titans Adopt NVIDIA DGX Cloud to Speed Success
Amgen, one of the world’s leading biotechnology companies, insurance technology leader CCC Intelligent Solutions (CCC), and digital-business-platform provider ServiceNow are among the first AI pioneers using DGX Cloud.
Amgen is using DGX Cloud with NVIDIA BioNeMo™ large language model software to accelerate drug discovery, including NVIDIA AI Enterprise software, which includes NVIDIA RAPIDS™ data science acceleration libraries.
“With NVIDIA DGX Cloud and NVIDIA BioNeMo, our researchers are able to focus on deeper biology instead of having to deal with AI infrastructure and set up ML engineering,” said Peter Grandsard, executive director of Research, Biologics Therapeutic Discovery, Center for Research Acceleration by Digital Innovation at Amgen. “The powerful computing and multi-node capabilities of DGX Cloud have enabled us to achieve 3x faster training of protein LLMs with BioNeMo and up to 100x faster post-training analysis with NVIDIA RAPIDS relative to alternative platforms.”
CCC, a leading cloud platform for the property and casualty insurance economy, is using DGX Cloud to speed and scale the development and training of its AI models. These models power the company’s innovative auto claims resolution solutions, helping to accelerate the intelligent automation of the industry and improve the claims experience for millions of business users and their consumers every day.
ServiceNow is using DGX Cloud with on-premises NVIDIA DGX supercomputers for flexible, scalable hybrid-cloud AI supercomputing that helps power its AI research on large language models, code generation, and causal analysis. ServiceNow also co-stewards the BigCode project, a responsible open-science generative AI initiative, which is trained on the Megatron-LM framework from NVIDIA.
Open a Browser to NVIDIA AI Supercomputing and Software
Enterprises manage and monitor DGX Cloud training workloads using NVIDIA Base Command™ Platform software, which provides a seamless user experience across DGX Cloud, as well as on-premises NVIDIA DGX supercomputers. Using Base Command Platform, customers can match their workloads to the right amount and type of DGX infrastructure needed for each job.
DGX Cloud includes NVIDIA AI Enterprise, the software layer of the NVIDIA AI platform, which provides end-to-end AI frameworks and pretrained models to accelerate data science pipelines and streamline the development and deployment of production AI. New pretrained models, optimized frameworks and accelerated data science software libraries, available in NVIDIA AI Enterprise 3.1 released today, give developers an additional jump-start to their AI projects.
Each instance of DGX Cloud features eight NVIDIA H100 or A100 80GB Tensor Core GPUs for a total of 640GB of GPU memory per node. A high-performance, low-latency fabric built with NVIDIA Networking ensures workloads can scale across clusters of interconnected systems, allowing multiple instances to act as one massive GPU to meet the performance requirements of advanced AI training. High-performance storage is integrated into DGX Cloud to provide a complete solution for AI supercomputing.
DGX Cloud features support from NVIDIA experts throughout the AI development pipeline. Customers can work directly with NVIDIA engineers to optimize their models and quickly resolve development challenges across a broad range of industry use cases.
DGX Cloud instances start at $36,999 per instance per month. Organizations can contact their NVIDIA Partner Network representative for additional details.
Watch Huang discuss NVIDIA DGX Cloud in his GTC keynote on demand, and tune in to the GTC panel with NVIDIA DGX Cloud pioneers.