A100 80GB PCIe | A100 80GB SXM | |
FP64 | 9.7 TFLOPS | |
FP64 Tensor Core | 19.5 TFLOPS | |
FP32 | 19.5 TFLOPS | |
Tensor Float 32 (TF32) | 156 TFLOPS | 312 TFLOPS* | |
BFLOAT16 Tensor Core | 312 TFLOPS | 624 TFLOPS* | |
FP16 Tensor Core | 312 TFLOPS | 624 TFLOPS* | |
INT8 Tensor Core | 624 TOPS | 1248 TOPS* |
GPU Memory | 80GB HBM2e | 80GB HBM2e |
GPU Memory Bandwidth | 1,935 GB/s | 2,039 GB/s |
Max Thermal Design Power (TDP) | 300W | 400W *** |
Multi-Instance GPU | Up to 7 MIGs @ 10GB | Up to 7 MIGs @ 10GB |
Form Factor | PCIe | SXM |
Dual-slot air-cooled or single-slot liquid-cooled | ||
Interconnect | NVIDIA® NVLink® Bridge | NVLink: 600 GB/s |
for 2 GPUs: 600 GB/s ** | PCIe Gen4: 64 GB/s | |
Server Options | Partner and NVIDIA-Certified Systems™ with 1-8 GPUs | NVIDIA HGX™ A100-Partner and NVIDIA-Certified Systems with 4,8, or 16 GPUs NVIDIA DGX™ A100 with 8 GPUs |
|