PRODUCTS: GPU

NVIDIAH200

 Supercharges generative AI as the first GPU with HBM3e, H200’s faster, larger memory fuels the acceleration of generative AI and LLMs while advancing scientific computing for HPC workloads.
H200
141 GB

MEMORY

HBM3e per GPU
4.8TB/s

MEMORY BANDWIDTH

 HBM3e per GPU
1.6X

GPT-3 175B inference

 Performance vs H100
1.9X

Llama2 70B Inference

 Performance vs H100 
2.0X

HPC Simulation

 Performance vs H100

NVIDIA H200 TENSOR CORE GPU

Key Features

 

 

NVIDIA H200 TENSOR CORE GPU

TECH SPECS

 

 

HBM3e Memory

The H200 is the world’s first GPU to feature HBM3e memory, delivering 4.8TB/s of bandwidth—a 43% boost over the H100—and expanding memory capacity to 141GB, nearly double the H100’s 80GB. This advancement dramatically improves data throughput for generative AI and high-performance computing (HPC) workloads. For large language models (LLMs) like GPT-3, the H200 delivers up to 18X the performance of the A100 and 11X faster performance compared to the H100.

Unmatched Performance

The H200 delivers 32 petaflops of FP8 deep learning performance, making it a powerful solution for the most demanding AI workloads. It also triples the FLOPS of double-precision Tensor Cores, providing 67 teraFLOPS of FP64 computing power for high-performance computing (HPC). For AI-accelerated HPC tasks, the H200 supports TF32 Tensor Core precision, enabling nearly one petaFLOP of single-precision matrix-multiply performance—without requiring any code modifications.

Advanced Architecture

Powered by NVIDIA Hopper™ architecture, the H200 is designed for continuous performance gains through ongoing software updates.

Compatibility

Fully compatible with existing HGX H100 systems, enabling seamless integration and performance upgrades without the need for infrastructure modifications.

Versatile Deployment

Designed for versatile deployment across data center environments, including on-premises, cloud, hybrid-cloud, and edge infrastructures.

GPU Memory

141GB

GPU Memory Bandwidth

4.8TB/s

FP8 Tensor Core Performance

4 PetaFLOPS

Form Factor

SXM | PCIe

Server Options

NVIDIA HGX H200 partner and NVIDIA-certified systems with 4 or 8 GPUs

Nvidia Enterprise 5.0

Included

PRODUCTS: ACCELERATED COMPUTING PLATFORM

NVIDIAHGX200

The NVIDIA HGX H200™️ AI supercomputing platform brings together the full power of NVIDIA GPUs, NVLink®️, NVIDIA networking, and fully optimized AI and high-performance computing (HPC) software stacks to provide the highest application performance and drive the fastest time to insights.
h200-tensor
8xH200

GPUs

8x NVIDIA H200 SXM
32 PFLOPS

Tensor Core

FP8/FP6 TENSOR CORE
900 GB/s

Bandwidth

NVSwitch GPU-GPU Bandwidth
7.2TB/s

Agg. Bandwidth

Total aggregate bandwidth 
NVIDIA HGX 200 COMPUTING PLATFORM

Key Features

 

 

NVIDIA HGX 200 COMPUTING PLATFORM

TECH SPECS

 

 

Scalable Configuration

Offered in both four- and eight-GPU configurations to accommodate a wide range of computing requirements.

Eight-Way Configuration

An eight-GPU HGX H200 system delivers more than 32 petaflops of FP8 deep learning performance and offers a total of 1.1TB of high-bandwidth memory.

Seamless Integration

Compatible with current HGX H100 systems, enabling effortless and cost-effective upgrades.

NVIDIA NVLink and NVSwitch

High-speed interconnects support the development of powerful scale-up servers, ensuring fast and efficient communication across multiple GPUs.

INT8 Tensor Core

32 POPS

FP16/BFLOAT16 Tensor Core

16 PFLOPS

TF32 Tensor Core

8 PFLOPS

FP32

540 TFLOPS

Nvidia Enterprise 5.0

Included

PRODUCTS: SOFTWARE

NVIDIA AI ENTERPRISE 5.0

NVIDIA®️ AI Enterprise is an all-encompassing, secure AI software platform designed to accelerate the data science pipeline and streamline the development and deployment of AI in production environments. As an end-to-end solution, it offers enterprises a robust, stable, and cloud-native platform packed with over 100 frameworks, pretrained models, and tools, covering a wide range of AI applications including generative AI, computer vision, and speech AI.
NV-AI-EP5
5x

Data Processing

Improve data processing time by up to 5 times
4x

Cheaper Operations

Reduces operational costs by 4 times
40x

Acceleration

Accelerate application performance up to 40 times
100+

Frameworks

Includes 100+ frameworks, pre-trained models, and tools
NVIDIA AI ENTERPRISE 5.0 SOFTWARE PLATFORM

Benefits for enterprises

 

 

NVIDIA AI ENTERPRISE 5.0 SOFTWARE PLATFORM

Top Use Cases

 

 

Secure + Stable Platform

Supports a wide range of AI workloads—including generative AI, computer vision, speech AI, and more—through a secure, stable, cloud-native platform featuring over 100 AI frameworks, pretrained models, and development tools.

Data preparation

Boost data processing speeds by up to 5× and cut operational costs by 4× using the NVIDIA RAPIDS Accelerator for Apache Spark.

AI Training

Build precise custom models in hours—not months—using the NVIDIA TAO Toolkit combined with pretrained models.

Optimization for Inference

Boost application inference performance up to 40× compared to CPU-only platforms using NVIDIA TensorRT.

Deployment at Scale

Streamline and enhance large-scale AI model deployment in production environments with NVIDIA Triton Inference Server.

LLM/Generative AI with Nemo

NV_NEMO

Extract, transform, and load(ETL)/Data Processing with RAPIDS Accelerator for Apache Spark

NV_ACCELERATORSPARK

Inference with TensorRT, Triton Inference Server and Triton Management Service

NV_TENSORRT

Speech and Translation AI with Riva

NV_RIVA

Cybersecurity with Morpheus

NV_MORPHEUS

Healthcare with MONAI and Parabricks

NV_MONAI