NVIDIA Quantum-2 InfiniBand Platform

Extreme performance for cloud-native supercomputing at any scale

Introducing QuantumX-800 for AI-Optimized Networking in Every Data Center

Record-Breaking Performance in Network Communications

NVIDIA Quantum-2, the seventh generation of the NVIDIA InfiniBand architecture, gives AI developers and scientific researchers the fastest networking performance and feature-sets available to take on the world’s most challenging problems. NVIDIA Quantum-2 empowers the world’s leading supercomputing data centers with software-defined networking, In-Network Computing, performance isolation, advanced acceleration engines, remote direct-memory access (RDMA), and the fastest speeds and feeds up to 400Gb/s.

Data Speed

Improved Performance

Improved TCO

Exascale Ready

Accelerated Deep Learning

Enhancing HPC and AI Supercomputers and Applications

Accelerated In-Network Computing

Today’s high-performance computing (HPC), AI, and hyperscale infrastructures require faster interconnects and more intelligent networks to analyze data and run complex simulations with greater speed and efficiency. NVIDIA Quantum-2 enhances and extends its In-Network Computing with preconfigured and programmable compute engines, such as the third generation of NVIDIA Scalable Hierarchical Aggregation and Reduction Protocol (SHARPv3)™, Message Passing Interface (MPI) Tag Matching, MPI All-to-All, and programmable engines, delivering the best cost per node and ROI.

Performance Isolation

The NVIDIA Quantum-2 InfiniBand platform provides innovative proactive monitoring and congestion management to deliver traffic isolations, nearly eliminating performance jitter, and ensuring predictive performance as if the application is being run on a dedicated system.

Cloud-Native Supercomputing

The NVIDIA Cloud-Native Supercomputing platform leverages the NVIDIA® BlueField® data processing unit (DPU) architecture with high-speed, low-latency NVIDIA Quantum-2 InfiniBand networking. The solution delivers bare-metal performance, user management and isolation, data protection, on-demand high performance computing (HPC), and AI services—simply and securely.

Delivering Data at the Speed of Light

Adapters

The NVIDIA ConnectX-7 InfiniBand adapter, with PCIe Gen4 and Gen5 support, is available in various form factors, delivering single or dual network ports at 400Gb/s.

The ConnectX-7 InfiniBand adapters include advanced In-Network Computing capabilities and additional programmable engines that enable both preprocessing data algorithms and offload application control paths to the network.

Switches

The NVIDIA Quantum-2 family of switches comprises 64 400Gb/s ports or 128 200Gb/s ports on physical 32 octal small form-factor (OSFP) connectors. The compact 1U switch design includes air-cooled and liquid-cooled versions that are either internally or externally managed.

The NVIDIA Quantum-2 family of switches delivers an aggregated 51.2 terabits per second (Tb/s) of bidirectional throughput with a capacity of more than 66.5 billion packets per second (bpps).

Transceivers and Cables

The NVIDIA Quantum-2 connectivity options provide maximum flexibility to build a topology of choice. They include a variety of transceivers and multi-fiber push-on connectors (MPOs), active copper cables (ACCs), and direct attached cables (DACs) with 1–2 and 1–4 splitter options.

Backward compatibility is also available to connect new 400Gb/s clusters to existing 200Gb/s or 100Gb/s-based infrastructures.

World-Leading Networking Performance, Scalability, and Efficiency

Performance

  • 400Gb/s bandwidth per port
  • 64 400Gb/s ports or 128 200Gb/s ports in a single switch
  • Over 66.5 billion packets per second (bidirectional) from a single NVIDIA Quantum-2 switch device

Breaking Our Own Records

  • 2X the bandwidth per port versus previous generation
  • 3X the switch radix versus previous generation
  • 4X MPI performance
  • 32X higher AI acceleration power per switch versus previous generation
  • Over one million 400Gb/s nodes in a four-switch-tier (three hops) DragonFly+ network, 6.5X higher than the previous generation
  • 7% reduction in data center power and space

Key Features

  • Full transport offload
  • RDMA, GPUDirect® RDMA, GPUDirect Storage
  • Programmable In-Network Computing engines
  • MPI All-to-All hardware acceleration
  • MPI Tag Matching hardware acceleration
  • NVIDIA SHARPv3
  • Advanced adaptive routing, congestion control, and QoS
  • Self-healing networking

NVIDIA Quantum-2 InfiniBand Platform