Magnum IO

Multi-GPU, Multi-Node Network and Storage IO Optimization Stack

Meeting the Bandwidth Demands of Compute-Intensive Workloads

GPUs provide the horsepower required by compute-intensive workloads, but their data consumption increases the demand for input/output (IO) bandwidth. NVIDIA Magnum IO is NVIDIA’s set of APIs that integrates computing, networking, file systems, and storage to maximize IO performance for multi-GPU, multi-node accelerated systems. It interfaces with CUDA-X libraries to accelerate IO across a broad range of workloads, from AI to visualization.

GPU-Optimized Networking and Storage IO Performance

Magnum IO integrates infrastructure elements to maximize storage and network I/O performance and functionality. Key benefits include:

Optimized IO Performance: It bypasses the CPU to enable direct IO between GPU memory and network storage.

System Balance and Utilization: It relieves CPU contention to create a more balanced GPU-accelerated system and delivers peak IO bandwidth with up to 10X fewer CPU cores.

Seamless Integration: It provides optimized implementation for current and future platforms, whether the data transfers are latency sensitive, bandwidth sensitive, or collectives.

IO Optimization Stack

Magnum IO includes innovative IO optimization technologies such as NCCL, NVIDIA® GPUDirect RDMA, and NVIDIA Fabric Manager. GPUDirect Storage is a key feature of the stack. It opens a direct data path between GPU memory and storage, avoiding the CPU altogether. This direct path can increase bandwidth, decrease latency, and decrease the utilization load on the CPU and GPU. Addressing the IO problem with Magnum IO drives toward a balanced system.

IO Optimization Stack

Learn more about the technologies powering Magnum IO.