In case you are a die-hard Nvidia loyalist, be able to pay a fortune to make use of its AI factories within the cloud.
Renting the GPU firm’s DGX Cloud, which is an all-inclusive AI supercomputer within the cloud, begins at $36,999 per occasion for a month.
The rental consists of entry to a cloud laptop with eight Nvidia H100 or A100 GPUs and 640GB of GPU reminiscence. The worth consists of the AI Enterprise software program to develop AI functions and enormous language fashions similar to BioNeMo.

“DGX Cloud has its personal pricing mannequin, so prospects pay Nvidia they usually can procure it by way of any of the cloud marketplaces primarily based on of the situation they select to devour it at, but it surely’s a service that’s priced by Nvidia, all inclusive,” mentioned Manuvir Das, vp for enterprise computing at Nvidia, throughout a briefing with press.
The DGX Cloud beginning value is near double that of $20,000 charged by Microsoft Azure for a fully-loaded A100 occasion with 96 CPU cores, 900GB of storage and eight A100 GPUs per 30 days.
Oracle is internet hosting DGX Cloud infrastructure in its RDMA Supercluster, which scales to 32,000 GPUs. Microsoft will launch DGX Cloud subsequent quarter, with Google Cloud’s implementation coming after that.
Prospects should pay a premium for the most recent {hardware}, however the integration of software program libraries and instruments could enchantment to enterprises and knowledge scientists.
Nvidia argues it offers one of the best accessible {hardware} for AI. Its GPUs are the cornerstone for high-performance and scientific computing.
However Nvidia’s proprietary {hardware} and software program is like utilizing the Apple iPhone – you’re getting one of the best {hardware}, however as soon as you’re locked in, it is going to be arduous to get out, and it’ll value some huge cash in its lifetime.
However paying a premium for Nvidia’s GPUs might convey long-term advantages. For instance, Microsoft is investing in Nvidia {hardware} and software program as a result of it presents value financial savings and bigger income alternatives by way of Bing with AI.
The idea of an AI manufacturing facility was floated by CEO Jensen Huang, who envisioned knowledge as uncooked materials, with the manufacturing facility turning it into usable knowledge or a complicated AI mannequin. Nvidia’s {hardware} and software program are the principle elements of the AI manufacturing facility.
“You simply present your job, level to your knowledge set and also you hit go and all the orchestration and all the things beneath is taken care of in DGX Cloud. Now the identical mannequin is on the market on infrastructure that’s hosted at a wide range of public clouds,” mentioned Manuvir Das, vp for enterprise computing at Nvidia, throughout a briefing with press.
Hundreds of thousands of individuals are utilizing ChatGPT-style fashions, which require high-end AI {hardware}, Das mentioned.
DGX Cloud furthers Nvidia’s purpose to promote its {hardware} and software program as a set. Nvidia’s is transferring into the software program subscription enterprise, which has an extended tail that entails promoting extra {hardware} so it will possibly generate extra software program income.
A software program interface, the Base Command Platform, will permit firms to handle and monitor DGX Cloud coaching workloads.
The Oracle Cloud has clusters of as much as 512 Nvidia GPUs, with a 200 gigabits-per-second RDMA community. The infrastructure helps a number of file methods together with Lustre and has 2 terabytes per second throughput.
Nvidia additionally introduced that extra firms had adopted its H100 GPU. Amazon is asserting their EC2 “UltraClusters” with P5 cases, which can be primarily based on the H100.
“These cases can scale as much as 20,000 GPUs utilizing their EFA expertise,” mentioned Ian Buck, vp of hyperscale and HPC computing at Nvidia in the course of the press briefing.
The EFA expertise refers to Elastic Cloth Adapter, which is a networking implementation orchestrated by Nitro, which is an all-purpose customized chip that handles networking, safety and knowledge processing.
Meta Platforms has begun the deployment of H100 methods in Grand Teton, the platform for the social media firm’s subsequent AI supercomputer.