Cloud and Data Center Solutions

Data Center Products

Accelerating the next wave of AI innovation.

Products

NVIDIA Vera Rubin NVL72

NVIDIA Vera Rubin NVL72 unifies 72 NVIDIA Rubin GPUs, 36 NVIDIA Vera CPUs, NVIDIA® ConnectX®-9 SuperNICs, and NVIDIA BlueField®-4 DPUs. It scales up intelligence in a rack-scale platform with the sixth-generation NVIDIA NVLink™ and NVLink Switch and scales out with NVIDIA Quantum-X800 InfiniBand and NVIDIA Spectrum-X™ Ethernet to power the AI industrial revolution at scale.

NVIDIA Groq 3 LPX

NVIDIA Groq 3 LPX™ is the inference accelerator for NVIDIA Vera Rubin, designed to meet the low-latency and large-context demands of agentic systems. NVIDIA Vera Rubin and LPX unite the extreme performance of NVIDIA Rubin GPUs and LPUs through a co-designed architecture. LPX features 256 LPUs with 128 gigabytes (GB) of SRAM, 40 petabytes per second (PB/s) of memory bandwidth, and 640 terabytes per second (TB/s) of scale-up bandwidth per rack.

NVIDIA DGX Vera Rubin NVL72

NVIDIA DGX™ Vera Rubin NVL72 provides enterprises with a turnkey, ready-to-deploy AI infrastructure solution built upon the NVIDIA Rubin platform, purpose-built to be deployed at scale to accelerate the most complex AI models.

NVIDIA HGX Rubin NVL8

The NVIDIA HGX™ platform brings together the full power of NVIDIA GPUs, NVIDIA NVLink, NVIDIA networking, and fully optimized AI and high-performance computing (HPC) software stacks to drive the highest application performance and the fastest time to insights for every data center.

NVIDIA DGX Rubin NVL8

NVIDIA DGX Rubin NVL8 is a liquid-cooled AI system powered by eight NVIDIA Rubin GPUs and sixth-generation NVLink, purpose-built to accelerate training, inference, and post-training for every AI workload.

NVIDIA Vera CPU

NVIDIA Vera is built for reinforcement learning and agentic AI, powering the code, tools, and data workflows that operate beyond the model. As the host CPU in GPU-accelerated systems, Vera pairs seamlessly with NVIDIA GPUs, directing data movement, managing memory, and orchestrating system control to keep AI pipelines running at full speed.

NVIDIA GB300 NVL72

The NVIDIA GB300 NVL72 connects 36 NVIDIA Grace™ CPUs and 72 NVIDIA Blackwell Ultra GPUs in a rack-scale design, supercharging inference, training, and data processing. The GB300 NVL72 is built for AI reasoning and delivers breakthrough performance, enabling large-scale reasoning applications to efficiently handle user requests at scale.

NVIDIA GB200 NVL72

The NVIDIA GB200 NVL72 connects 36 NVIDIA Grace CPUs and 72 NVIDIA Blackwell GPUs in a rack-scale design, supercharging generative AI, data processing, and high-performance computing. The GB200 NVL72 is a liquid-cooled, rack-scale solution that boasts a 72-GPU NVLink domain, powered by NVIDIA Mission Control™, that acts as a single massive GPU and delivers 30x faster inference for trillion-parameter large language models (LLMs).

NVIDIA RTX PRO 4500 Blackwell Server Edition

The NVIDIA RTX PRO™ 4500 Blackwell Server Edition GPU, based on the revolutionary NVIDIA Blackwell architecture, is an energy-efficient, multi-workload accelerator designed to deliver breakthrough performance across a broad range of enterprise workloads, from AI inference, data science, and processing to video and high-end visual computing.

NVIDIA RTX PRO 6000 Blackwell Server Edition

The NVIDIA RTX PRO 6000 Blackwell Server Edition GPU delivers a powerful combination of AI and visual computing capabilities to accelerate enterprise data center workloads. Equipped with 96 GB of ultra-fast GDDR7 memory, the RTX PRO 6000 Blackwell provides unparalleled performance and flexibility to accelerate a broad range of use cases, from agentic AI, physical AI, and scientific computing to rendering, 3D graphics, and video.

Architectures

NVIDIA Vera Rubin

NVIDIA Vera Rubin is built for the next generation of AI, supercharging agentic AI and AI reasoning. With extreme co-design across seven chips, Vera Rubin delivers flexible deployment options while slashing training and inference token costs, dramatically lowering the total cost of AI deployment.

NVIDIA Blackwell

NVIDIA Blackwell is defining the next chapter in generative AI and accelerated computing with unparalleled performance, efficiency, and scale. NVIDIA Blackwell features six transformative technologies that unlock breakthroughs in data processing, electronic design automation, computer-aided engineering, and quantum computing.

NVIDIA MGX

NVIDIA MGX™ is a modular reference architecture for accelerated computing that supports hundreds of GPU, DPU, CPU, storage, and networking combinations for AI, high-performance computing, and NVIDIA Omniverse™ workloads.

NVIDIA STX

By leveraging the NVIDIA Vera Rubin architecture, BlueField-4, and Spectrum-X networking, the NVIDIA STX reference architecture provides the foundation for building a universal data engine that accelerates the full AI lifecycle, from training and analytics to real-time agentic inference.

Platforms

NVIDIA DGX

Built from the ground up for enterprise AI, the NVIDIA DGX platform incorporates the best of NVIDIA software, infrastructure, and expertise in a modern, unified AI development solution that spans from the cloud to on-premises data centers.

NVIDIA HGX

The NVIDIA HGX platform brings together the full power of NVIDIA GPUs, NVLink, networking, and fully optimized AI and high-performance computing software stacks to drive the highest application performance and the fastest time to insights for every data center.

NVIDIA Spectrum-X Ethernet

The NVIDIA Spectrum-X Ethernet networking platform delivers the highest performance for AI. It connects compute fabrics within the data center and scales across multiple AI data centers with Spectrum-XGS Ethernet technology to form massive AI super-factories capable of gigascale intelligence.

NVIDIA RTX PRO

NVIDIA RTX PRO™ Servers from global system partners leverage the breakthrough performance and energy efficiency of the NVIDIA Blackwell architecture, enabling enterprises to build AI factories and accelerate a wide range of enterprise workloads, from agentic AI and LLM inference to industrial AI and digital twins.

Networking Purpose-Built for AI Factories

Explore the full-stack networking fabric for gigascale AI factories, from NVIDIA Spectrum-X Ethernet to NVIDIA BlueField DPUs.

FAQs

The NVIDIA Vera Rubin platform is built for the age of agentic AI and reasoning, engineered to master multi-step problem-solving and massive long-context workflows at scale. By eliminating critical bottlenecks in communication and memory movement, the platform supercharges inference to deliver more tokens per watt and lower cost per token versus the NVIDIA Blackwell architecture generation.

NVIDIA Vera Rubin NVL72 features our third-generation MGX NVL72 rack, which preserves the same rack footprint as NVIDIA Blackwell while introducing new mechanical, cooling, and power enhancements.

NVIDIA DGX Vera Rubin NVL72 is the next iteration of NVIDIA's rack-scale system architecture. These systems are built using the NVIDIA Rubin architecture, combining 72 NVIDIA Rubin GPUs and 36 NVIDIA Vera CPUs with sixth-generation NVLink to create a massive shared memory space capable of accelerating the world's most complex models. DGX Vera Rubin NVL72 also leverages the latest innovations in NVIDIA networking, equipped with ConnectX-9 SuperNICs, BlueField-4 DPUs, and Spectrum–X Ethernet.

Get Started

Ready to Get Started?

NVIDIA data center solutions are available through select NVIDIA Partner Network (NPN) partners. Explore flexible, affordable options for accessing the latest NVIDIA data center technologies through our partners.

Get the Latest Data Center News

Sign up for enterprise news, announcements, and more from NVIDIA.