NVIDIA DGX A100

The Universal System for AI Infrastructure

The World’s First AI System Built on NVIDIA A100

NVIDIA DGX A100 is the universal system for all AI workloads, offering unprecedented compute density, performance, and flexibility in the world’s first 5 petaFLOPS AI system. Featuring the NVIDIA A100 Tensor Core GPU, DGX A100 enables enterprises to consolidate training, inference, and analytics into a unified, easy-to-deploy AI infrastructure that includes direct access to NVIDIA AI experts.

Essential Building Block of the AI Data Center

The Universal System for Every AI Workload

The Universal System for Every AI Workload

NVIDIA DGX A100 is the universal system for all AI infrastructure, from analytics to training to inference. DGX A100 is powered by NVIDIA Base Command - the operating system of the accelerated datacenter. It sets a new bar for compute density, packing 5 petaFLOPS of AI performance into a 6U form factor, replacing legacy infrastructure silos with one platform for every AI workload.

DGXperts: Integrated Access to AI Expertise

DGXperts: Integrated Access to AI Expertise

NVIDIA DGXperts are a global team of 20,000+ AI-fluent professionals who have built a wealth of experience over the last decade to help you maximize the value of your DGX investment.

Fastest Time To Solution

Fastest Time To Solution

NVIDIA DGX A100 is the world’s first AI system built on the NVIDIA A100 Tensor Core GPU. Integrating eight A100 GPUs with 320GB or 640GB of GPU memory, the system provides unprecedented acceleration and is fully optimized for NVIDIA CUDA-X software and the end-to-end NVIDIA data center solution stack.

data center-scalability-icon

Unmatched Data Center Scalability

NVIDIA DGX A100 features NVIDIA® ConnectX®-7 InfiniBand/Ethernet network adapters with 500 gigabytes per second (GB/s) of peak bidirectional bandwidth. This is one of the many features that make DGX A100 the foundational building block for large AI clusters such as NVIDIA DGX SuperPOD, the enterprise blueprint for scalable AI infrastructure.

A Simpler and Faster Way to Tackle AI

A Simpler and Faster Way to Tackle AI

The NVIDIA AI Starter Kit provides everything your team needs—from a world-class AI platform, to optimized software and tools, to consultative services—to get your AI initiatives up and running quickly. Don't lose time and money building an AI platform. Plug in and power up in a day, get use cases defined in a week, and start productizing models sooner.

  • For Developers
  • For IT Managers
  • For Business Leaders
Solve Bigger, More Complex Data Science Problems Faster

Solve Bigger, More Complex Data Science Problems Faster

Benefits

  • Eliminate tedious set up and testing with ready-to-run, optimized AI software.
  • Create better models faster with unmatched performance that enables more iterations.
  • Avoid time lost on systems integration and software engineering.
Deploy Infrastructure and Operationalize AI at Scale

Deploy Infrastructure and Operationalize AI at Scale

Benefits

  • Experience simplified infrastructure design and capacity planning with one system for all AI workloads.
  • Achieve the highest compute density and performance in the smallest footprint.
  • Take advantage of built-in security at every layer.

 

Shorten Time to Insights and Accelerate ROI of AI

Shorten Time to Insights and Accelerate ROI of AI

Benefits

  • Increase data scientist productivity and eliminate non-value-added effort.
  • Accelerate the development cycle, from concept to production.
  • Remove roadblocks with advice from DGXperts.
  •  

Game Changing Performance

Training

DLRM Training

Up to 3X Higher Throughput
for AI Training on Largest Models

Inference

RNN-T Inference: Single Stream

Up to 1.25X Higher Throughput
for AI Inference

Data Analytics

Big Data Analytics Benchmark

Up to 83X Higher Throughput than CPU,
2X Higher Throughput than DGX A100 320GB

Explore the Powerful Components of DGX A100

18x NVIDIA A100 GPUs with up to 640 GB Total GPU Memory
12 NVLinks/GPU, 600 GB/s GPU-to-GPU Bi-directonal Bandwidth

2 6x NVIDIA NVSwitches
4.8 TB/s Bi-directional Bandwidth, 2X More than Previous Generation NVSwitch

3 Up to 10X NVIDIA Connectx-7 200 Gb/s Network Interface
500GB/s Peak Bidirectional Bandwidth

4 Dual 64-Core AMD CPUs and 2 TB System Memory
3.2X More Cores to Power the Most Intensive AI Jobs

5 30 TB Gen4 NVME SSD
50 GB/s Peak Bandwidth, 2X Faster than Gen3 NVME SSDs

Data Center

Take a deep dive into the new NVIDIA DGX A100.

The Technology Inside NVIDIA DGX A100

  • A100 GPU
  • Multi-Instance GPU
  • NVLink & NVSwitch
  • ConnectX-7
  • Optimized Software Stack
  • Security

NVIDIA A100 Tensor Core GPU

The NVIDIA A100 Tensor Core GPU delivers unprecedented acceleration for AI, data analytics, and high-performance computing (HPC) to tackle the world’s toughest computing challenges. With third-generation NVIDIA Tensor Cores providing a huge performance boost, the A100 GPU can efficiently scale up to the thousands or, with Multi-Instance GPU, be allocated as seven smaller, dedicated instances to accelerate workloads of all sizes.

NVIDIA A100 Tensor Core GPU

Multi-Instance GPU (MIG)

With MIG, the eight A100 GPUs in DGX A100 can be configured into as many as 56 GPU instances, each fully isolated with their own high-bandwidth memory, cache, and compute cores. This allows administrators to right-size GPUs with guaranteed quality of service (QoS) for multiple workloads.

Multi Instance GPU

Next-Generation NVLink and NVSwitch

The third generation of NVIDIA® NVLink in DGX A100 doubles the GPU-to-GPU direct bandwidth to 600 gigabytes per second (GB/s), almost 10X higher than PCIe Gen4. DGX A100 also features next-generation NVIDIA NVSwitch™, which is 2X times faster than the previous generation.

Next-Generation NVLink and NVSwitch

NVIDIA ConnectX-7 200Gb/s InfiniBand/Ethernet

DGX A100 features the NVIDIA ConnectX-7 InfiniBand and VPI (Infiniband or Ethernet) adapters, each running at 200 gigabits per second (Gb/s) to create a high-speed fabric for large-scale AI workloads. DGX A100 systems are also available with ConnectX-6 adapters.

NVIDIA ConnectX-7 200Gb/s InfiniBand/Ethernet

Optimized Software Stack

DGX A100 integrates a tested and optimized DGX software stack, including an AI-tuned base operating system, all necessary system software, and GPU-accelerated applications, pre-trained models, and more from NGC.

DGX Software Stack

Built-In Security

DGX A100 delivers the most robust security posture for AI deployments, with a multi-layered approach stretching across the baseboard management controller (BMC), CPU board, GPU board, self-encrypted drives, and secure boot.

Built-In Security

Hear more about NVIDIA DGX A100 from our DGXperts at GTC

Learn about enterprise-grade support for NVIDIA DGX systems.

NVIDIA DGX SuperPOD

Leadership-Class AI Infrastructure

Purchase NVIDIA DGX Today

Learn About NVIDIA DGX Systems