← Back to Library
Networking Provider: InfiniBand Trade Association

InfiniBand

InfiniBand is a high-performance networking technology providing RDMA capabilities with 200-400 Gb/s speeds and sub-microsecond latency. Dominant in AI/HPC clusters for GPU-to-GPU communication, enabling distributed training at scale. Features HDR (200 Gb/s) and NDR (400 Gb/s) standards, NVIDIA/Mellanox ecosystem dominance. Essential for multi-node GPU training, offering 10× lower latency than Ethernet. Used in 90% of top supercomputers and AI training clusters.

InfiniBand
networking hpc high-performance

Overview

InfiniBand is a high-performance networking architecture designed for HPC and AI workloads requiring maximum bandwidth and minimum latency. Unlike traditional Ethernet, InfiniBand uses Remote Direct Memory Access (RDMA) to enable direct memory-to-memory transfers between servers without CPU involvement, eliminating network protocol overhead. This makes it essential for distributed AI training where hundreds or thousands of GPUs must exchange gradients and model parameters with microsecond-level latency.

Current standards include HDR (High Data Rate) at 200 Gb/s per port and NDR (Next Data Rate) at 400 Gb/s per port. InfiniBand dominates the AI training market with NVIDIA/Mellanox providing the majority of hardware and software. It's used in 90% of the TOP500 supercomputers and virtually all large-scale AI training clusters from OpenAI, Google DeepMind, Meta, and Anthropic. For distributed training, InfiniBand provides 5-10× better performance than Ethernet alternatives like RoCE (RDMA over Converged Ethernet).

Key Features

  • **RDMA (Remote Direct Memory Access)**: Zero-copy networking, bypass OS kernel for minimal latency
  • **Ultra-low latency**: Sub-microsecond latency (0.5-0.7 μs), 10× lower than Ethernet
  • **High bandwidth**: HDR 200 Gb/s, NDR 400 Gb/s per port (25-50 GB/s actual throughput)
  • **GPU Direct RDMA**: Direct GPU-to-GPU communication across nodes without CPU involvement
  • **Adaptive routing**: Automatic traffic load balancing across multiple paths
  • **Hardware offload**: Protocol processing in NIC hardware, freeing CPU resources
  • **Fabric management**: OpenSM subnet manager for topology discovery and path setup
  • **MPI optimization**: Native support for Message Passing Interface, ideal for distributed training

Use Cases

  • **Distributed AI training**: Multi-node GPU clusters for training LLMs, diffusion models (GPT, Llama, DALL-E scale)
  • **High-performance computing**: Molecular dynamics, climate modeling, computational fluid dynamics
  • **Supercomputing**: 90% of TOP500 supercomputers use InfiniBand for node interconnect
  • **Storage area networks**: Parallel file systems (Lustre, GPFS) requiring high IOPS and low latency
  • **Database clusters**: In-memory databases (SAP HANA) with RDMA for fast data access
  • **Financial trading**: Ultra-low latency networks for algorithmic trading systems
  • **AI inference clusters**: Real-time inference for large models requiring multi-GPU deployment
  • **Scientific research**: Large-scale simulations, genomics, particle physics requiring massive compute

Technical Specifications

InfiniBand generations: EDR (Enhanced Data Rate, 100 Gb/s, legacy), HDR (High Data Rate, 200 Gb/s, current), NDR (Next Data Rate, 400 Gb/s, latest). Link widths: 1×, 4×, 8×, 12× lanes (NDR 12× = 4.8 Tb/s aggregate). Latency: 0.5-0.7 microseconds hardware latency, 1-2 microseconds MPI latency. Cable types: Copper (DAC) up to 5m, active optical cables (AOC) up to 100m, fiber optics for longer distances.

Network topologies: Fat-tree (most common for AI clusters), dragonfly, hypercube. Switch configurations: Leaf-spine for scalability, blocking ratios (1:1 non-blocking optimal). Subnet management via OpenSM or vendor tools (NVIDIA UFM). Performance: 90-95% link utilization achievable with proper tuning. RDMA protocols: IB Verbs API, RDMA CM (Connection Manager), OpenFabrics stack. GPU integration: NVIDIA GPUDirect RDMA eliminates CPU bounce buffers for 2-5× faster multi-node training.

Pricing

InfiniBand hardware pricing (approximate): HDR 200G adapter (NIC): $800-1,200 per port. NDR 400G adapter: $1,500-2,500 per port. HDR switches: $15K-50K for 36-port leaf switch, $100K-200K for 40-port spine switch. NDR switches: $200K-500K for enterprise spine switches. Cables: $50-200 for copper DAC, $200-500 for AOC, $500-1,500 for fiber. Total cluster cost: $2K-5K per node for adapters/cables, $50K-500K for switches depending on scale. Cloud options: AWS EC2 EFA (Elastic Fabric Adapter, InfiniBand-like) included, Azure InfiniBand on HPC VMs, CoreWeave InfiniBand clusters.

Code Example

# Distributed PyTorch training with InfiniBand/NCCL\nimport torch\nimport torch.distributed as dist\nfrom torch.nn.parallel import DistributedDataParallel as DDP\n\n# Initialize process group with NCCL backend (uses InfiniBand via RDMA)\ndist.init_process_group(\n    backend='nccl',\n    init_method='tcp://master_node:23456',\n    world_size=32,  # 32 GPUs across 4 nodes\n    rank=int(os.environ['RANK'])\n)\n\n# Create model and wrap with DDP for multi-GPU training\nmodel = YourLargeModel().cuda()\nmodel = DDP(model, device_ids=[local_rank])\n\n# Training loop - gradients sync across InfiniBand\nfor batch in dataloader:\n    outputs = model(batch)\n    loss = criterion(outputs, labels)\n    loss.backward()  # NCCL AllReduce over InfiniBand\n    optimizer.step()\n\n# Environment variables for InfiniBand optimization\n# export NCCL_IB_DISABLE=0  # Enable InfiniBand\n# export NCCL_IB_HCA=mlx5_0  # Mellanox adapter\n# export NCCL_IB_GID_INDEX=3  # RoCE GID\n# export NCCL_DEBUG=INFO  # Debug output\n\n# MPI configuration for multi-node\n# mpirun -np 32 \\\n#   -H node1:8,node2:8,node3:8,node4:8 \\\n#   --map-by ppr:8:node \\\n#   -x NCCL_IB_DISABLE=0 \\\n#   python train.py

Professional Integration Services by 21medien

21medien offers comprehensive integration services for InfiniBand, including API integration, workflow automation, performance optimization, custom development, architecture consulting, and training programs. Our experienced team helps businesses leverage InfiniBand for production applications with enterprise-grade reliability, security, and support. We provide end-to-end solutions from initial consultation and proof of concept through full-scale deployment, optimization, and ongoing maintenance. Our services include custom feature development, third-party integrations, migration assistance, performance tuning, and dedicated technical support. Schedule a free consultation through our contact page to discuss your specific requirements and explore how InfiniBand can transform your AI capabilities and accelerate your digital transformation initiatives.

Resources

Official website: https://www.infinibandta.org