Visit your regional NVIDIA website for local content, pricing, and where to buy partners specific to your country.
Bring end-to-end high-performance networking to scientific computing, AI, and cloud data centers.
Complex workloads demand ultra-fast processing of high-resolution simulations, extreme-size datasets, and highly parallelized algorithms. As these needs continue to grow, NVIDIA Quantum InfiniBand—the world’s only fully offloadable, In-Network Computing platform—provides dramatic leaps in performance to achieve faster time to discovery with less cost and complexity.
NVIDIA® ConnectX® InfiniBand host channel adapters (HCAs) provide ultra-low latency, extreme throughput, and innovative NVIDIA In-Network Computing engines to deliver the acceleration, scalability, and feature-rich technology needed for today's modern workloads.
The NVIDIA BlueField® DPUs combine powerful computing, high-speed networking, and extensive programmability to deliver software-defined, hardware-accelerated solutions for the most demanding workloads. From accelerated AI and scientific computing to cloud-native supercomputing, BlueField redefines what’s possible.
InfiniBand switch systems deliver the highest performance and port density available. Innovative capabilities such as NVIDIA Scalable Hierarchical Aggregation and Reduction Protocol (SHARP)™ and advanced management features such as self-healing network capabilities, quality of service, enhanced virtual lane mapping, and NVIDIA In-Network Computing acceleration engines provide a performance boost for industrial, AI, and scientific applications.
InfiniBand systems provide the highest scalability and subnet isolation using InfiniBand routers, and InfiniBand to Ethernet gateway systems. The latter is used to enable a scalable and efficient way to connect InfiniBand data centers to Ethernet infrastructures.
NVIDIA MetroX® long-haul systems can seamlessly connect remote InfiniBand data centers, storage, and other InfiniBand platforms. They can extend the reach of InfiniBand up to 40 kilometers, enabling native InfiniBand connectivity between remote data centers or between data center and remote storage infrastructures for high availability and disaster recovery.
LinkX® cables and transceivers are designed to maximize the performance of HPC networks, requiring high-bandwidth, low-latency, highly reliable connections between InfiniBand elements.
NVIDIA Scalable Hierarchical Aggregation and Reduction Protocol (SHARP)™ offloads collective communication operations to the switch network, decreasing the amount of data traversing the network, reducing the time of Message Passing Interface (MPI) operations, and increasing data center efficiency.
NVIDIA InfiniBand with self-healing network capabilities overcomes link failures, enabling network recovery 5,000X faster than any other software-based solution. These capabilities take advantage of the intelligence built into the latest generation of InfiniBand switches.
InfiniBand is the only high-performance interconnect solution with proven quality-of-service capabilities, including advanced congestion control and adaptive routing, resulting in unmatched network efficiency.
InfiniBand offers centralized management and supports any topology, including Fat Tree, Hypercubes, multi-dimensional Torus, and Dragonfly+. Routing algorithms optimize performance when designing a topology for particular application communication patterns.
OFED from OpenFabrics Alliance has been collaboratively developed and tested by high-performance input/output (IO) manufacturers. NVIDIA MLNX_OFED is an NVIDIA-tested version of OFED.
The NVIDIA HPC-X® is a comprehensive MPI and SHMEM/PGAS software suite. HPC-X leverages InfiniBand In-Network Computing and acceleration engines to optimize research and industry applications.
The NVIDIA Unified Fabric Manager (UFM®) platform empowers data center administrators to efficiently provision, monitor, manage, and proactively troubleshoot their InfiniBand network infrastructure.
NVIDIA Magnum IO™ utilizes network IO, In-Network Computing, storage, and IO management to simplify and speed up data movement, access, and management for multi-GPU, multi-node systems.