NVIDIA Gives Insant Access to AI Supercomputer From a BrowserNVIDIA Launches DGX Cloud, Giving Every Enterprise Instant Access to AI Supercomputer From a BrowserOracle Cloud Infrastructure First to Run NVIDIA AI Supercomputing Instances; Microsoft Azure, Google Cloud and Others to Host DGX Cloud SoonNVIDIA announced NVIDIA DGX™ Cloud, an AI supercomputing service that gives enterprises immediate access to the infrastructure and software needed to train advanced models for generative AI and other groundbreaking applications. DGX Cloud provides dedicated clusters of NVIDIA DGX AI supercomputing, paired with NVIDIA AI software. The service makes it possible for every enterprise to access its own AI supercomputer using a simple web browser, removing the complexity of acquiring, deploying and managing on-premises infrastructure. Enterprises rent DGX Cloud clusters on a monthly basis, which ensures they can quickly and easily scale the development of large, multi-node training workloads without having to wait for accelerated computing resources that are often in high demand. “We are at the iPhone moment of AI. Startups are racing to build disruptive products and business models, and incumbents are looking to respond,” said Jensen Huang, founder and CEO of NVIDIA. “DGX Cloud gives customers instant access to NVIDIA AI supercomputing in global-scale clouds.” NVIDIA is partnering with leading cloud service providers to host DGX Cloud infrastructure, starting with Oracle Cloud Infrastructure (OCI). Its OCI Supercluster provides a purpose-built RDMA network, bare-metal compute and high-performance local and block storage that can scale to superclusters of over 32,000 GPUs. Microsoft Azure is expected to begin hosting DGX Cloud next quarter, and the service will soon expand to Google Cloud and more. Industry Titans Adopt NVIDIA DGX Cloud to
Speed Success Amgen is using DGX Cloud with NVIDIA BioNeMo™ large language model software to accelerate drug discovery, including NVIDIA AI Enterprise software, which includes NVIDIA RAPIDS™ data science acceleration libraries. “With NVIDIA DGX Cloud and NVIDIA BioNeMo, our researchers are able to focus on deeper biology instead of having to deal with AI infrastructure and set up ML engineering,” said Peter Grandsard, executive director of Research, Biologics Therapeutic Discovery, Center for Research Acceleration by Digital Innovation at Amgen. “The powerful computing and multi-node capabilities of DGX Cloud have enabled us to achieve 3x faster training of protein LLMs with BioNeMo and up to 100x faster post-training analysis with NVIDIA RAPIDS relative to alternative platforms.” CCC a leading cloud platform for the property and casualty insurance economy, is using DGX Cloud to speed and scale the development and training of its AI models. These models power the company’s innovative auto claims resolution solutions, helping to accelerate the intelligent automation of the industry and improve the claims experience for millions of business users and their consumers every day. ServiceNow is using DGX Cloud with on-premises NVIDIA DGX supercomputers for flexible, scalable hybrid-cloud AI supercomputing that helps power its AI research on large language models, code generation, and causal analysis. ServiceNow also co-stewards the BigCode project, a responsible open-science generative AI initiative, which is trained on the Megatron-LM framework from NVIDIA. Open a Browser to NVIDIA AI Supercomputing and
Software DGX Cloud includes NVIDIA AI Enterprise, the software layer of the NVIDIA AI platform, which provides end-to-end AI frameworks and pretrained models to accelerate data science pipelines and streamline the development and deployment of production AI. New pretrained models, optimized frameworks and accelerated data science software libraries, available in NVIDIA AI Enterprise 3.1 released today, give developers an additional jump-start to their AI projects. Each instance of DGX Cloud features eight NVIDIA H100 or A100 80GB Tensor Core GPUs for a total of 640GB of GPU memory per node. A high-performance, low-latency fabric built with NVIDIA Networking ensures workloads can scale across clusters of interconnected systems, allowing multiple instances to act as one massive GPU to meet the performance requirements of advanced AI training. High-performance storage is integrated into DGX Cloud to provide a complete solution for AI supercomputing. DGX Cloud features support from NVIDIA experts throughout the AI development pipeline. Customers can work directly with NVIDIA engineers to optimize their models and quickly resolve development challenges across a broad range of industry use cases. Source: NVIDIA media announcement |