Dual-slot PCIe form factor of the H200 with the same 141 GB HBM3e memory and 4.8 TB/s bandwidth as the SXM variant. Lower TDP (600 W vs 700 W) makes it a drop-in upgrade for existing PCIe servers, with NVLink Bridge support between paired cards.
VRAM
141 GB
Memory
HBM3e
Bandwidth
4800 GB/s
TDP
600W
Large Language Models
Training and inference for models like GPT-4, Llama 70B+
Deep Learning Training
High-performance training for neural networks
Distributed Training
Multi-node training with fast interconnects
High-Throughput Inference
Optimized for batched inference workloads
Compute throughput shown with 2:4 structured sparsity. ~84% of H200 SXM compute due to lower TDP.
Estimates based on INT8 quantization. Actual fit depends on framework and batch size.
Added Apr 30, 2026
Last updated: Apr 30, 2026
From model selection to production, one platform, no fragmentation.