English

An In-Depth Look at NVIDIA NDR 400G InfiniBand

Posted on May 21, 2024 by
257

The relentless demand for faster data processing and transfer speeds in high-performance computing (HPC) and cloud computing has led to groundbreaking innovations in networking technology. At the forefront is NVIDIA Mellanox with its NDR InfiniBand(IB), bringing an impressive 400Gb/s capability. This article delves into the InfiniBand technology standard, its key components, and the performance enhancements delivered by this stellar advancement.

InfiniBand Architecture

InfiniBand employs two queue pair extraction techniques that allow applications to directly transfer data from the adapter to application memory (referred to as RDMA) and vice versa. Compared to operations in the TCP/IP protocol, IB utilizes Single Data Copy (SDP) to directly copy data from the network card to the user's application space, bypassing involvement of the core memory. This zero-copy approach achieves maximum protocol throughput during large-scale data processing. The InfiniBand protocol adopts a layered structure, with each layer being independent and providing services to the layer above it. The physical layer defines how bit signals form symbols on the wire and then how symbols form frames, data symbols, and data padding. It provides detailed specifications for constructing efficient data packets and signaling protocols. You can read the article InfiniBand Network and Architecture Overview to learn more.

InfiniBand Basic Components

The InfiniBand network topology consists of four categories of components:

  • Host Channel Adapter (HCA): It acts as a bridge between the memory controller and the Target Channel Adapter (TCA).

  • Target Channel Adapter (TCA): It packages digital signals from I/O devices (e.g., network cards, SCSI controllers) and sends them to the HCA.

  • InfiniBand links: They are the fiber connections between HCAs and TCAs. The IB architecture allows hardware vendors to connect TCAs and HCAs using 1, 4, or 12 fiber cables.

  • Switches and routers: Both HCAs and TCAs essentially act as host adapters and feature programmable Direct Memory Access (DMA) engines with certain protection capabilities.

Rate Evolution of InfiniBand

In high-concurrency and high-performance computing applications where customers demand exceptional bandwidth and low latency, IB networking emerges as a suitable solution. Currently, InfiniBand supports various bandwidth modes, and future plans include even higher modes. The following table presents the bandwidth modes using 4X as an example:

Mode Bandwidth
SRD (Single Data Rate) 8 Gb/s
DDR (Double Data Rate) 16 Gb/s
QDR (Quad Data Rate) 32 Gb/s
FDR (Fourteen Data Rate) 56 Gb/s
EDR (Enhanced Data Rate) 100 Gb/s
HDR (High Data Rate) 200 Gb/s
NDR (Next Data Rate) 400 Gb/s+

Overview of NVIDIA NDR 400Gb/s InfiniBand

Comparisons with Previous InfiniBand Generations

When comparing NDR 400Gb/s InfiniBand with previous generations such as EDR and HDR, it becomes evident that the former offers substantial advantages. The increased bandwidth and reduced latency of NDR 400Gb/s IB outperform its predecessors, enabling faster and more efficient data transfers. Moreover, compatibility and interoperability considerations ensure a smooth transition for organizations already utilizing IB technology.

Performance Enhancements of NVIDIA NDR 400G InfiniBand

NVIDIA Mellanox NDR 400Gb/s InfiniBand boosts performance in multiple ways:

  • 1. Higher Bandwidth: It delivers faster data transfer, improving communication efficiency in HPC clusters.

  • 2. Reduced Latency: It minimizes communication delays, enabling rapid data transmission and responsiveness.

  • 3. Enhanced Scalability: It supports large-scale HPC environments, handling increased data volumes and facilitating parallel processing across clusters.

  • 4. Improved Efficiency: NDR IB optimizes data transfer, minimizing overhead and maximizing network resource utilization for higher system efficiency and application performance.

  • 5. Advanced Features: It introduces enhanced error detection, congestion control, and supports advanced network topologies, ensuring better reliability and stability.

  • 6. Future-Proofing: NDR IB's high data transfer rate accommodates growing workloads and emerging technologies without the need for extensive infrastructure upgrades, providing a technology roadmap for future HPC requirements.

NVIDIA 400G InfiniBand Products Available from FS

Categories Product Ports/Speeds PCIe Interface Features
Switch MQM9700-NS2F 64 x NDR 400G
32 OSFP Connectors
/ Managed
MQM9790-NS2F 64 x NDR 400G
32 OSFP Connectors
/ Unmanaged
Adapter MCX75510AAS-NEAT 1 x 400G NDR
Single-Port OSFP
PCIe 5.0 x 16 InfiniBand
MCX75310AAS-NEAT 1 x 400G NDR
Single-Port OSFP
PCIe 5.0 x 16 InfiniBand & Ethernet
Secure Boot
MCX75310AAC-NEAT 1 x 400G NDR
Single-Port OSFP
PCIe 5.0 x 16 InfiniBand & Ethernet
Crypto and Secure
Boot
MCX715105AS-WEAT 1 x 400G NDR
Single-Port QSFP112
PCIe 5.0 x 16 InfiniBand & Ethernet
Secure Boot
400G NDR InfiniBand DAC & AOC / / OSFP
OSFP to 2x QSFP56
OSFP to OSFP
Modules / / SMF/MMF
800G NDR InfiniBand DAC / /

OSFP
OSFP to OSFP
OSFP to 2x OSFP
OSFP to 2x QSFP112
OSFP to 4x OSFP
OSFP to 4x QSFP112

Modules / / SMF/MMF

Conclusion

NVIDIA NDR 400G InfiniBand technology sets a new benchmark for high-speed networking. Its performance enhancements, including increased bandwidth, reduced latency, scalability, and reliable data transmission, empower organizations across industries to achieve new levels of productivity and efficiency, driving the next generation of computing.

You might be interested in

Knowledge
See profile for Howard.
Howard
InfiniBand Network and Architecture Overview
Dec 30, 2023
2.4k
Knowledge
See profile for Howard.
Howard
An Overview of NVIDIA NVLink
Jan 29, 2024
14.5k
Knowledge
Knowledge
Knowledge
Knowledge
See profile for Sheldon.
Sheldon
Decoding OLT, ONU, ONT, and ODN in PON Network
Mar 14, 2023
402.4k
Knowledge
See profile for Irving.
Irving
What's the Difference? Hub vs Switch vs Router
Dec 17, 2021
373.0k
Knowledge
See profile for Sheldon.
Sheldon
What Is SFP Port of Gigabit Switch?
Jan 6, 2023
348.8k
Knowledge
Knowledge
See profile for Migelle.
Migelle
PoE vs PoE+ vs PoE++ Switch: How to Choose?
May 30, 2024
431.3k