
✓ GPU memory: 24GB HBM2
✓ GPU memory bandwidth: 933GB/s
✓ Interconnect: PCIe Gen4: 64GB/s, Third-gen NVLINK: 200GB/s**
✓ Form factor: Dual-slot, full-height, full-length (FHFL)
✓ Max thermal design power (TDP): 165W
✓ Multi-Instance GPU (MIG): 4 GPU instances @ 6GB each, 2 GPU instances @ 12GB each, 1 GPU instance @ 24GB
✓ Virtual GPU (vGPU) software support: NVIDIA AI Enterprise, NVIDIA Virtual Compute Server
✓ Warranty: 2 year manufacturer repair or replace
Ships in 7 days from payment. No returns or cancellations. All sales final. For bulk inquiries, consult a live chat agent or call our toll-free number.
Bring accelerated performance to every enterprise workload with NVIDIA A30 Tensor Core GPUs. With NVIDIA Ampere architecture Tensor Cores and Multi-Instance GPU (MIG), it delivers speedups securely across diverse workloads, including AI inference at scale and high-performance computing (HPC) applications. By combining fast memory bandwidth and low-power consumption in a PCIe form factor—optimal for mainstream servers—A30 enables an elastic data center and delivers maximum value for enterprises.
FP64 | 5.2 teraFLOPS | |
FP64 Tensor Core | 10.3 teraFLOPS | |
FP32 | 10.3 teraFLOPS | |
TF32 Tensor Core | 82 teraFLOPS | 165 teraFLOPS* | |
BFLOAT16 Tensor Core | 165 teraFLOPS | 330 teraFLOPS* | |
FP16 Tensor Core | 165 teraFLOPS | 330 teraFLOPS* | |
INT8 Tensor Core | 330 TOPS | 661 TOPS* | |
INT4 Tensor Core | 661 TOPS | 1321 TOPS* | |
Media engines | 1 optical flow accelerator (OFA) 1 JPEG decoder (NVJPEG) 4 video decoders (NVDEC) | |
GPU memory | 24GB HBM2 | |
GPU memory bandwidth | 933GB/s | |
Interconnect | PCIe Gen4: 64GB/s Third-gen NVLINK: 200GB/s** | |
Form factor | Dual-slot, full-height, full-length (FHFL) | |
Max thermal design power (TDP) | 165W | |
Multi-Instance GPU (MIG) | 4 GPU instances @ 6GB each 2 GPU instances @ 12GB each 1 GPU instance @ 24GB | |
Virtual GPU (vGPU) software support | NVIDIA AI Enterprise NVIDIA Virtual Compute Server |
No reviews available.