Nvidia DGX A100 Datasheet
Nvidia DGX A100 Datasheet
Nvidia DGX A100 Datasheet
survive, but to thrive in challenging times. However, the enterprise requires a GPUs 8x NVIDIA A100 8x NVIDIA A100
80 GB GPUs 40 GB GPUs
platform for AI infrastructure that improves upon traditional approaches, which
GPU Memory 640 GB total 320 GB total
historically involved slow compute architectures that were siloed by analytics,
Performance 5 petaFLOPS AI
training, and inference workloads. The old approach created complexity,
10 petaOPS INT8
drove up costs, constrained speed of scale, and was not ready for modern AI.
NVIDIA 6
Enterprises, developers, data scientists, and researchers need a new platform NVSwitches
that unifies all AI workloads, simplifying infrastructure and accelerating ROI. System Power 6.5 kW max
Usage
The Universal System for Every AI Workload CPU Dual AMD Rome 7742, 128 cores total,
2.25 GHz (base), 3.4 GHz (max boost)
NVIDIA DGX ™ A100 is the universal system for all AI workloads—from analytics System Memory 2 TB 1 TB
to training to inference. DGX A100 sets a new bar for compute density, packing Networking 8x Single- 8x Single-
Port Mellanox Port Mellanox
5 petaFLOPS of AI performance into a 6U form factor, replacing legacy ConnectX-6 VPI ConnectX-6 VPI
compute infrastructure with a single, unified system. DGX A100 also offers 200Gb/s HDR 200Gb/s HDR
InfiniBand InfiniBand
the unprecedented ability to deliver fine-grained allocation of computing power,
2x Dual-Port 1x Dual-Port
using the Multi-Instance GPU (MIG) capability in the NVIDIA A100 Tensor Core Mellanox Mellanox
ConnectX-6 VPI ConnectX-6 VPI
GPU, which enables administrators to assign resources that are right-sized for
10/25/50/100/200 10/25/50/100/200
specific workloads. Available with up to 640 gigabytes (GB) of total GPU memory, Gb/s Ethernet Gb/s Ethernet
which increases performance in large-scale training jobs up to 3X and doubles Storage OS: 2x 1.92 TB M.2 OS: 2x 1.92TB M.2
NVME drives NVME drives
the size of MIG instances, DGX A100 can tackle the largest and most complex
Internal Storage: Internal Storage:
jobs, along with the simplest and smallest. Running the DGX software stack 30 TB (8x 3.84 TB) 15 TB (4x 3.84 TB)
U.2 NVMe drives U.2 NVMe drives
with optimized software from NGC, the combination of dense compute power
Software Ubuntu Linux OS
and complete workload flexibility make DGX A100 an ideal choice for both single
Also supports:
node deployments and large scale Slurm and Kubernetes clusters deployed Red Hat Enterprise Linux
with NVIDIA DeepOps. CentOS
Direct Access to NVIDIA DGXperts Packaged System 359.7 lbs (163.16 kgs) max
Weight
NVIDIA DGX A100 is more than a server. It’s a complete hardware and software System Height: 10.4 in (264.0 mm)
platform built upon the knowledge gained from the world’s largest DGX proving Dimensions Width: 19.0 in (482.3 mm) max
ground—NVIDIA DGX SATURNV—and backed by thousands of DGXperts at Length: 35.3 in (897.1 mm) max
NVIDIA. DGXperts are AI-fluent practitioners who offer prescriptive guidance Operating 5–30 ºC (41–86 ºF)
Temperature
and design expertise to help fastrack AI transformation. They've built a wealth Range
1.25X
1X
NVIDIA DGX A100 features eight NVIDIA A100 Tensor Core 1X
1X
GPUs, which deliver unmatched acceleration, and is 1X
fully optimized for NVIDIA CUDA-X ™ software and the 0.7X
floating operations per second (FLOPS) for AI compared DLRM on HugeCTR framework, precision = FP16 | 1x DGX
MLPerf 0.7 RNN-T measured with (1/7) MIG slices. Frame-
work: TensorRT 7.2, dataset = LibriSpeech, precision = FP16.
to the previous generation. Best of all, no code changes A100 640GB batch size = 48 | 2x DGX A100 320GB batch
size = 32 | 1x DGX-2 (16x V100 32GB) batch size = 32.
Speedups Normalized to Number of GPUs.
are required to achieve this speedup. And when using
NVIDIA’s automatic mixed precision with FP16, A100
offers an additional 2X boost to performance with just one
additional line of code. Up to 83X Higher Throughput than CPU, 2X Higher Throughput than
DGX A100 320GB on Big Data Analytics Benchmark
The A100 80GB GPU doubles the high-bandwidth memory
from 40 GB (HBM) to 80GB (HBM2e) and increases GPU 90X
for Enterprise Big data analytics benchmark | 30 analytical retail queries, ETL, ML, NLP on 10TB dataset | CPU: 19x Intel Xeon Gold
6252 2.10 GHz, Hadoop | 16x DGX-1 (8x V100 32GB each), RAPIDS/Dask | 12x DGX A100 320GB and 6x DGX A100
640GB, RAPIDS/Dask/BlazingSQL. Speedups Normalized to Number of GPUs.
NVIDIA DGX A100 delivers the most robust security posture
for your AI enterprise, with a multi-layered approach that
secures all major hardware and software components. InfiniBand adapters for clustering and up to two dual-port ConnectX-6
Stretching across the baseboard management controller VPI Ethernet adapters for storage and networking, all capable of 200 Gb/s.
(BMC), CPU board, GPU board, self-encrypted drives, The combination of massive GPU-accelerated compute with state-of-
and secure boot, DGX A100 has security built in, allowing the-art networking hardware and software optimizations means DGX
IT to focus on operationalizing AI rather than spending A100 can scale to hundreds or thousands of nodes to meet the biggest
time on threat assessment and mitigation. challenges, such as conversational AI and large-scale image classification.