// performance, scalability, and security

NVIDIA H100

Tap into unprecedented performance, scalability, and security for every workload with the NVIDIA H100 Tensor Core GPU. With the NVIDIA NVLink Switch System, up to 256 H100 GPUs can be connected to accelerate exascale workloads. The GPU also includes a dedicated Transformer Engine to solve trillion-parameter language models.

The H100’s combined technology innovations can speed up large language models (LLMs) by an incredible 30X over the previous generation to deliver industry-leading conversational AI.

// For the highest workload

Powered by the NVIDIA Hopper Architecture

NVIDIA H100 GPUs feature fourth-generation Tensor Cores and the Transformer Engine with FP8 precision, further extending NVIDIA’s market-leading AI leadership with up to 4X faster training and an incredible 30X inference speedup on large language models. For high-performance computing (HPC) applications, H100 triples the floating-point operations per second (FLOPS) of FP64 and adds dynamic programming (DPX) instructions to deliver up to 7X higher performance. With second-generation Multi-Instance GPU (MIG), built-in NVIDIA confidential computing, and NVIDIA NVLink Switch System, H100 securely accelerates all workloads for every data center from enterprise to exascale.

Accelerate Next-Generation Workloads

  • Generative AI
  • Large language model (LLM) training and inference
  • HPC simulations
  • Extreme performance
GB HBM2e

GPU Memory

80
TFLOPS

Single-Precision Performance

67
TFLOPS

AI Tensor Performance (FP8)

3958

// Still not sure what is the best GPU for you? We are ready to assist you.

NEED A CONSULTATION?