NVIDIA GB300 NVL72

NVIDIA GB300 NVL72

Built for the age of AI reasoning.

Available Now

Overview

Designed for AI Reasoning Performance

The NVIDIA GB300 NVL72 features a fully liquid-cooled, rack-scale architecture that integrates 72 NVIDIA Blackwell Ultra GPUs and 36 Arm®-based NVIDIA Grace™ CPUs into a single platform. The system delivers 1.5x more dense FP4 Tensor Core FLOPS and 2x higher attention performance compared to NVIDIA Blackwell GPUs. It’s purpose-built for test-time scaling inference and AI reasoning tasks. AI factories accelerated by the GB300 NVL72—leveraging NVIDIA Quantum-X800 InfiniBand or Spectrum-X™ Ethernet, ConnectX-8 SuperNIC™s, and NVIDIA Mission Control management—deliver up to a 50x overall increase in AI factory output performance compared to NVIDIA Hopper-based platforms.

Performance

Increase AI Factory Output Performance by 50x

DeepSeek-R1 ISL = 32K, OSL = 8K, GB300 NVL72 with FP4 Dynamo disaggregation. H100 with FP8 In-flight batching. Projected performance subject to change.

Experience next-level AI reasoning performance with the NVIDIA GB300 NVL72 platform. Compared to Hopper, the GB300 NVL72 delivers an impressive 10x boost in user responsiveness (TPS per user) and a 5x improvement in throughput (TPS per megawatt (MW)). Together, these advancements translate into a remarkable 50x leap in overall AI factory output.

Accelerating Real-Time Video Generation by 30x

GB300 NVL72 introduces cutting-edge capabilities for diffusion-based video generation models. A single five-second video generation sequence processes 4 million tokens, requiring nearly 90 seconds to generate on NVIDIA Hopper GPUs. The Blackwell Ultra platform enables real-time video generation from world foundation models, such as NVIDIA Cosmos™, providing a 30x performance improvement versus Hopper. This allows the creation of customized, photo-realistic, temporally and spatially stable video for physical AI applications.

Projected performance subject to change. Relative performance represented as frames per second per GPU. 5-second video generation using Cosmos-1.0-Diffusion-7B-Video2World 720p 60 FPS.

Features

Technological Breakthroughs

AI Reasoning Inference

Test-time scaling and AI reasoning increase the compute necessary to achieve quality of service and maximum throughput. NVIDIA Blackwell Ultra’s Tensor Cores are supercharged with 2x the attention-layer acceleration and 1.5x more AI compute floating-point operations per second (FLOPS) compared to NVIDIA Blackwell GPUs.

High-Capacity HBM3E Architecture

Larger memory capacity allows for larger batch sizing and maximum throughput performance. NVIDIA Blackwell Ultra GPUs offer 1.5x larger HBM3E memory in combination with added AI compute compared with its predecessor, boosting AI reasoning throughput for the largest context lengths.

NVIDIA Blackwell Architecture

The NVIDIA Blackwell architecture delivers groundbreaking advancements in accelerated computing, powering a new era of unparalleled performance, efficiency, and scale.

NVIDIA ConnectX-8 SuperNIC

The NVIDIA ConnectX-8 SuperNIC’s input/output (IO) module hosts two ConnectX-8 devices, providing 800 gigabits per second (Gb/s) of network connectivity for each GPU in the NVIDIA GB300 NVL72. This delivers best-in-class remote direct-memory access (RDMA) capabilities with either NVIDIA Quantum-X800 InfiniBand or Spectrum-X Ethernet networking platforms, enabling peak AI workload efficiency.

NVIDIA Grace CPU

The NVIDIA Grace CPU is a breakthrough processor designed for modern data center workloads. It provides outstanding performance and memory bandwidth with 2x the energy efficiency of today’s leading server processors.

Fifth-Generation NVIDIA NVLink

Unlocking the full potential of accelerated computing requires seamless communication between every GPU. The fifth generation of NVIDIA NVLink™ is a scale–up interconnect that unleashes accelerated performance for AI reasoning models.

Automate the Essentials With NVIDIA Mission Control

NVIDIA Mission Control powers every aspect of NVIDIA GB300 NVL72 AI factory operations, from orchestrating workloads across the 72-GPU NVLink domain to integration with facilities. It brings instant agility for inference and training while providing full-stack intelligence for infrastructure resilience. Mission Control lets every enterprise run AI with hyperscale-grade efficiency, accelerating AI experimentation.

NVIDIA DGX Station

Develop and run cutting‑edge AI locally with NVIDIA DGX Station™, powered by the NVIDIA GB300 Grace Blackwell Ultra Desktop Superchip. Accelerate large-scale training and inferencing workloads and easily scale them from deskside to data center.

Specifications

NVIDIA GB300 NVL72¹

Configuration 72 NVIDIA Blackwell Ultra GPUs, 36 NVIDIA Grace CPUs
NVLink Bandwidth 130 TB/s
Fast Memory 37 TB
GPU Memory | Bandwidth 20 TB | Up to 576 TB/s
CPU Memory | Bandwidth 17 TB LPDDR5X | 14 TB/s
CPU Core Count 2,592 Arm Neoverse V2 cores
FP4 Tensor Core 1440 | 10802 PFLOPS
FP8/FP6 Tensor Core 720 PFLOPS
INT8 Tensor Core 24 POPS
FP16/BF16 Tensor Core 360 PFLOPS
TF32 Tensor Core 180 PFLOPS
FP32 6 PFLOPS
FP64 / FP64 Tensor Core 100 TFLOPS

Resources

Learn More About Grace Blackwell

Get Started

Stay Informed

Reach out to sales to get more information on the NVIDIA GB300 NVL72.

Stay Up to Date on NVIDIA News

Sign up for the latest news, updates, and more from NVIDIA.