TESLA A30

Bring accelerated performance to every enterprise workload with NVIDIA A30 Tensor Core GPUs. With NVIDIA Ampere architecture Tensor Cores and Multi-Instance GPU (MIG), it delivers speedups securely across diverse workloads, including AI inference at scale and high-performance computing (HPC) applications. By combining fast memory bandwidth and low-power consumption in a PCIe form factor—optimal for mainstream servers—A30 enables an elastic data center and delivers maximum value for enterprises.

HD Gallery

  • DESCRIPTION
  • SPECIFICATION
  • BIOS / DRIVERS UPDATE
  • REQUIREMENT
►NVIDIA Ampere Architecture CUDA® Cores ►Third-Generation NVIDIA NVLink®
►Second-Generation RT Cores ►Virtualization-Ready
►Third-Generation Tensor Cores ►PCI Express Gen 4
►48GB of GPU Memory ►Data Center Efficiency and Security

FP64

5.2 teraFLOPS

FP64 Tensor Core

10.3 teraFLOPS

FP32

10.3 teraFLOPS

TF32 Tensor Core

82 teraFLOPS | 165 teraFLOPS*

BFLOAT16 Tensor Core

165 teraFLOPS | 330 teraFLOPS*

FP16 Tensor Core

165 teraFLOPS | 330 teraFLOPS*

INT8 Tensor Core

330 TOPS | 661 TOPS*

INT4 Tensor Core

661 TOPS | 1321 TOPS*

Media engines

1 optical flow accelerator (OFA)
1 JPEG decoder (NVJPEG)
4 video decoders (NVDEC)

GPU memory

24GB HBM2

GPU memory bandwidth

933GB/s

Interconnect

PCIe Gen4: 64GB/s
Third-gen NVLINK: 200GB/s**

Form factors

Dual-slot, full-height, full-length (FHFL)

Max thermal design power (TDP)

165W

Multi-Instance GPU (MIG)

4 GPU instances @ 6GB each
2 GPU instances @ 12GB each
1 GPU instance @ 24GB

Virtual GPU (vGPU) software support

NVIDIA AI Enterprise for VMware
NVIDIA Virtual Compute Server