NDR 400G InfiniBand Architecture

Delivers Extreme Performance for the Exascale AI Era

InfiniBand Networking Solutions

NVIDIA® Mellanox® InfiniBand interconnect brings high-speed, extremely low-latency scalable solutions incorporating In-Network Computing technology that performs data algorithms on the network, enabling the era of “data-centric” data centers. The InfiniBand technology enables supercomputer, Artificial Intelligence (AI) and cloud data centers to operate at any scale, while reducing operational costs and infrastructure complexity.

InfiniBand Adapters - Accelerating AI, Data Science, and HPC Workloads

InfiniBand Adapters - Accelerating AI, Data Science, and HPC Workloads

InfiniBand Host Channel Adapters (HCAs) leveraging faster speeds and innovative hardware accelerations, enable extremely low-latency, and advanced application acceleration engines for high-performance computing, artificial intelligence, machine learning, cloud, storage, databases, and embedded applications, reducing cost per operation and increasing overall ROI.

Programmable DPU

Programmable DPU

NVIDIA BlueField® Data Processing Unit (DPU) with advanced software and programmability, provides data-centers with levels of performance, security and functionality for innovative networking and I/O acceleration.

InfiniBand Switches - Unrivaled Performance and Economics at Scale

InfiniBand Switches - Unrivaled Performance and Economics at Scale

InfiniBand switch systems deliver the highest performance and port density available. Innovative capabilities such as Scalable Hierarchical Aggregation and Reduction Protocol (SHARP)and Self-Healing Interconnect Enhancement for Intelligent Datacenters (SHIELD), as well as novel quality of service (QoS), congestion control and adaptive routing provides the highest overall system performance, scalability and best network utilization.

Routers & Gateway Systems to Bridge InfiniBand to Ethernet

Routers & Gateway Systems to Bridge InfiniBand to Ethernet

InfiniBand systems provide the highest scalability and subnet isolation using InfiniBand Routers, InfiniBand long-reach connections (NVIDIA Mellanox MetroX®-2), and InfiniBand to Ethernet gateway systems (NVIDIA Mellanox Skyway). The latter are used to enable a scalable and efficient way to connect the high-performance, low-latency InfiniBand data center to external Ethernet infrastructures in case of need (e.g. Ethernet storage).

LinkX InfiniBand Cables and Transceivers

LinkX InfiniBand Cables and Transceivers

NVIDIA Mellanox LinkX® cables and transceivers are designed to maximize the performance of High Performance Computing networks, requiring high-bandwidth, low-latency, highly reliable connections between InfiniBand elements.

InfiniBand Enhanced Capabilities

In-Network Computing

In-Network Computing

NVIDIA Mellanox Scalable Hierarchical Aggregation and Reduction Protocol (SHARP) offloads collective communications operations from the CPU to the switch network. This innovative approach decreases the amount of data traversing the network and dramatically reduces the MPI operations time.

Self Healing Network

Self Healing Network

In HPC, clusters depend upon a high-speed and reliable interconnect. NVIDIA Mellanox InfiniBand with self-healing network capabilities, overcomes link failures, enabling network recovery 5000x faster than any other software-based solution. The self-healing networking capabilities take advantage of the intelligence already built into the latest generation of InfiniBand switches.

Quality of Service

Quality of Service

InfiniBand is the only high-performance interconnect solution with proven quality-of-service capabilities, including advanced congestion control and adaptive routing resulting in unmatched network efficiency.

Network Topologies

Network Topologies

InfiniBand has complete centralized management and can support any topology. The most popular topologies include Fat Tree, Hypercubes, multi-dimensional Torus and Dragonfly+. Optimized routing algorithms provide optimized performance when designing a topology for particular application communication patterns.

Software for Optimal Performance

MLNX_OFED

OFED from OpenFabrics Alliance (www.openfabrics.org) has been hardened through collaborative development and testing by major high performance I/O vendors. NVIDIA MLNX_OFED is an NVIDIA tested and packaged version of OFED.

HPC-X

The NVIDIA Mellanox HPC-X® ScalableHPC Toolkit is a comprehensive MPI and SHMEM/ PGAS software suite. HPC-X enables the rapid deployment and delivery of maximum application performance without the complexity and costs of licensed third-party tools and libraries.

UFM

NVIDIA Mellanox UFM® (Unified Fabric Management) platforms combine enhanced, real-time network telemetry with AI-powered cyber Intelligence and analytics, to realize higher utilization of fabric resources and a competitive advantage, while reducing OPEX.

MAGNUM IO

Magnum IO utilizes network IO, in-network compute, storage and IO management to simplify and speed up data movement, access, and management for multi-GPU, multi-node systems. Magnum IO enables NVIDIA GPU and NVIDIA networking hardware topologies to achieve optimal throughput and low latency.

Configuration Tool

Academy Online Courses

Ready to Purchase