// Acceleration at Every Scale

NVIDIA A100

NVIDIA A100 Tensor Core GPU delivers unprecedented acceleration at every scale to power the world’s highest-performing elastic data centers for AI, data analytics, and HPC. Powered by the NVIDIA Ampere Architecture, A100 is the engine of the NVIDIA data center platform. A100 provides up to 20X higher performance over the prior generation and can be partitioned into seven GPU instances to dynamically adjust to shifting demands.

The A100 80GB debuts the world’s fastest memory bandwidth at over 2 terabytes per second (TB/s) to run the largest models and datasets.

// Incredible Performance Across Workloads

Powered by the NVIDIA Ampere Architecture

NVIDIA A100 introduces double precision Tensor Cores to deliver the biggest leap in HPC performance since the introduction of GPUs. Combined with 80GB of the fastest GPU memory, researchers can reduce a 10-hour, double-precision simulation to under four hours on A100. HPC applications can also leverage TF32 to achieve up to 11X higher throughput for single-precision, dense matrix-multiply operations.

For the HPC applications with the largest datasets, A100 80GB’s additional memory delivers up to a 2X throughput increase with Quantum Espresso, a materials simulation. This massive memory and unprecedented memory bandwidth makes the A100 80GB the ideal platform for next-generation workloads.

Accelerate Next-Generation Workloads

  • Generative AI
  • Large language model (LLM) training and inference
  • HPC simulations
  • Extreme performance
GB HBM2e

GPU Memory

80
TFLOPS

Single-Precision Performance

19.5
TFLOPS

AI Tensor Performance (FP16)

624

// Still not sure what is the best GPU for you? We are ready to assist you.

NEED A CONSULTATION?