The V100 pioneered Tensor Cores and mixed-precision training with 32GB HBM2 memory. A landmark GPU that defined modern AI infrastructure, still deployed for many workloads.
VRAM
32 GB
Memory
HBM2
Bandwidth
900 GB/s
TDP
300W
Smaller Language Models
Inference for 7B-13B parameter models
Enterprise Deployment
Designed for 24/7 datacenter operations
Previous gen flagship
Estimates based on INT8 quantization. Actual fit depends on framework and batch size.
Added Jan 25, 2026
Last updated: Jan 25, 2026
Explore models, compare pricing and benchmarks, and right-size your infrastructure — all in one place.