NVIDIA A2 Tensor Core GPU
Entry-level GPU that brings NVIDIA AI to any server.
Manufacturer Part Number: 900-2G179-0020-001
Features and Benefits:
The NVIDIA A2 Tensor Core GPU provides entry-level inference with low power, a small footprint, and high performance for NVIDIA AI at the edge. Featuring a low-profile PCIe Gen4 card and a low 40-60W configurable thermal design power (TDP) capability, the A2 brings versatile inference acceleration to any server for deployment at scale.
Specifications:
Peak FP32 |
4.5 TF |
TF32 Tensor Core |
9 TF | 18 TF¹ |
BFLOAT16 Tensor Core |
18 TF | 36 TF¹ |
Peak FP16 Tensor Core |
18 TF | 36 TF¹ |
Peak INT8 Tensor Core |
36 TOPS | 72 TOPS¹ |
Peak INT4 Tensor Core |
72 TOPS | 144 TOPS¹ |
RT Cores |
10 |
Media engines |
1 video encoder
2 video decoders (includes AV1 decode) |
GPU memory |
16GB GDDR6 |
GPU memory bandwidth |
200GB/s |
Interconnect |
PCIe Gen4 x8 |
Form factor |
1-slot, low-profile PCIe |
Max thermal design power (TDP) |
40–60W (configurable) |
Virtual GPU (vGPU) software support² |
NVIDIA Virtual PC (vPC), NVIDIA Virtual Applications (vApps), NVIDIA RTX Virtual Workstation (vWS), NVIDIA AI Enterprise, NVIDIA Virtual Compute Server (vCS) |
1 With sparsity
2 Supported in future vGPU release
point performance 2.91 Tflops (GPU Boost Clocks)
1.87 Tflops (Base Clocks) 1.66 Tflops (GPU Boost Clocks)
1.43 Tflops (Base Clocks)
Peak single precision floating
point performance 8.74 Tflops (GPU Boost Clocks)
5.6 Tflops (Base Clocks) 5 Tflops (GPU Boost Clocks)
4.29 Tflops (Base Clocks)
Memory bandwidth (ECC off)² 480 GB/sec (240 GB/sec per GPU) 288 GB/sec
Memory size (GDDR5) 24 GB (12GB per GPU) 12 GB
CUDA cores 4992 ( 2496 per GPU) 2880 - See more at: http://www.nvidia.com/object/tesla-servers.html#sthash.ZmsPP43F.dpuf
This product is special order, which takes longer to process and cannot be returned.