Today, NVIDIA announced NVIDIA DGX supercomputer powered by NVIDIA GH200 Grace Hopper Superchips and the NVIDIA NVLink Switch System. This supercomputer can be used to develop models for generative AI language applications, recommendation systems, data analytics and more. The NVIDIA DGX GH200 combines 256 GH200 superchips using NVIDIA NVLink Switch System to perform as a single GPU. The NVIDIA DGX GH200 will deliver 1 exaflop of performance and 144 terabytes of shared memory.
NVIDIA revealed that Google Cloud, Meta and Microsoft are among the first expected to gain access to this new supercomputer. NVIDIA will also offer DGX GH200 design as a blueprint to cloud service providers and other big tech companies allowing them to customize it for their infrastructure.
Highlights of NVIDIA DGX GH200 supercomputer:
- GH200 superchips eliminate the need for a traditional CPU-to-GPU PCIe connection by combining an Arm-based NVIDIA Grace CPU with an NVIDIA H100 Tensor Core GPU in the same package, using NVIDIA NVLink-C2C chip interconnects. This increases the bandwidth between GPU and CPU by 7x compared with the latest PCIe technology, slashes interconnect power consumption by more than 5x, and provides a 600GB Hopper architecture GPU building block for DGX GH200 supercomputers.
- DGX GH200 is the first supercomputer to pair Grace Hopper Superchips with the NVIDIA NVLink Switch System, a new interconnect that enables all GPUs in a DGX GH200 system to work together as one.
- The DGX GH200 architecture provides 48x more NVLink bandwidth than the previous generation, delivering the power of a massive AI supercomputer with the simplicity of programming a single GPU.
The new NVIDIA DGX GH200 supercomputers are expected to be available by the end of the year.