世界上最强大的GPU NVIDIA H200 Tensor Core GPU 具有改变游戏规则的性能和内存功能,可增强生成式 AI 和高性能计算 (HPC) 工作负载。作为首款采用 HBM3e 的 GPU,H200 更大更快的内存可加速生成式 AI 和大型语言模型 (LLM),同时推进 HPC 工作负载的科学计算。
规格 NVIDIA H200 | Tensor Core GPU 构成因素 H200 SXM¹ FP64 34 万亿次浮点运算 FP64 Tensor Core 67 万亿次浮点运算 FP32 67 万亿次浮点运算 TF32 Tensor Core 989 万亿次浮点运算² BFLOAT16 Tensor Core 1,979 TFLOPS² FP16 Tensor Core 1,979 TFLOPS² FP8 Tensor Core 3,958 TFLOPS² INT8 Tensor Core 3,958 TFLOPS² 显存 141GB GPU显存带宽 4.8TB/s 解码器 7 NVDEC 7 JPEG 最大热设计功率 (TDP) Up to 700W (configurable) 多实例 GPU Up to 7 MIGs @16.5GB each Form Factor SXM Interconnect NVIDIA NVLink®: 900GB/s PCIe Gen5: 128GB/s Server Options NVIDIA HGX™ H200 partner and NVIDIA-Certified Systems™ with 4 or 8 GPUs NVIDIA AI Enterprise Add-on |
NVIDIA H200 Tensor Core GPU
The world’s most powerful GPU for supercharging
AI and HPC workloads.
Higher Performance and Larger, Faster Memory
The NVIDIA H200 Tensor Core GPU supercharges generative AI and high-
performance computing (HPC) workloads with game-changing performance and
memory capabilities.
Based on the NVIDIA Hopper™ architecture, the NVIDIA H200 is the first GPU to
offer 141 gigabytes (GB) of HBM3e memory at 4.8 terabytes per second (TB/s)—that’s
nearly double the capacity of the NVIDIA H100 Tensor Core GPU with 1.4X more
memory bandwidth. The H200’s larger and faster memory accelerates generative AI
and large language models, while advancing scientific computing for HPC workloads
with better energy efficiency and lower total cost of ownership.
Unlock Insights With High-Performance LLM Inference
In the ever-evolving landscape of AI, businesses rely on large language models to
address a diverse range of inference needs. An AI inference accelerator must deliver
the highest throughput at the lowest TCO when deployed at scale for a massive
user base.
The H200 doubles inference performance compared to H100 GPUs when handling
large language models such as Llama2 70B.
Preliminary measured performance, subject to change.
Llama2 13B: ISL 128, OSL 2K | Throughput | H100 1x GPU BS 64 | H200 1x GPU BS 128
GPT-3 175B: ISL 80, OSL 200 | x8 H100 GPUs BS 64 | x8 H200 GPUs BS 128
Llama2 70B: ISL 2K, OSL 128 | Throughput | H100 1x GPU BS 8 | H200 1x GPU BS 32.
Datasheet