AI Hardware , Enterprise GPUs
Instock

Nvidia HGX H20 Enterprise 96GB

SKU:Nvidia H20
0 out of 5 (0)

Applications of H20

1. AI Inference & Large Language Models (LLMs)

  • Optimized for large AI models such as ChatGPT, Gemini, and Claude.
  • Designed for fast, efficient inference in cloud environments.
  • Reduces power consumption while maintaining high AI compute performance.

2. Cloud Computing & AI SaaS Services

  • Ideal for deployment on AWS, Google Cloud, Alibaba Cloud, and other cloud platforms.
  • Supports AI-based speech recognition, machine translation, and virtual assistants.
  • Provides a scalable, cost-effective AI infrastructure.

3. Medical AI (Medical Imaging & Genomic Analysis)

  • Enhances medical imaging recognition (CT/MRI analysis).
  • Accelerates protein folding prediction (AlphaFold) and genetic sequencing.
  • Reduces processing times for AI-driven diagnostics.
     
Get this product for
vipera
Get it in 10 days
vipera
Will be delivered to your location via DHL or UPS. Ask an agent if import tariffs apply.
This is a preorder. All sales final. No returns or cancellations. For bulk inquiries, consult a live chat agent or call our toll-free number.
NVIDIA HGX H20: Pinnacle of Hopper Architecture

Unveiling NVIDIA's HGX H20, a powerhouse driven by the cutting-edge Hopper architecture. With 96 GB of HBM3 memory and a 4.0 TB/s memory bandwidth, this GPU delivers unparalleled performance in AI applications. Diverse Tensor Cores, including INT8, FPB, BF16, and FP16, contribute to up to 296 TFLOPS, complemented by an additional 74 TFLOPS from the TF32 Tensor Core. Beyond raw power, the HGX H20 supports Multi-Instance GPU (MIG) technology, offering versatility with up to seven instances for optimized workload distribution. Enhanced by a 60 MB L2 Cache and a media engine featuring 7 NVDEC and 7 NVJPEG units, this GPU is designed for efficiency and robust multimedia processing. With a 400 W power consumption and a form factor tailored for eight-way HGX configurations, the HGX H20 marks a paradigm shift in GPU processing.

NVIDIA H20: The Next-Generation AI Inference GPU

The H20 is built on the Hopper architecture and features 14,592 CUDA cores. It integrates Tensor Cores optimized for AI workloads and supports the Transformer Engine, enabling highly efficient deep learning acceleration.

For memory, the H20 is equipped with 96GB of HBM3 memory with an ultra-high bandwidth of 4.0TB/s, significantly improving data transfer speeds. It supports NVLink for multi-GPU interconnect and uses the PCIe 5.0 interface.

The power consumption (TDP) of H20 is only 350W, making it much more energy-efficient than the 700W power draw of the H100 while maintaining strong AI compute capabilities. In FP16 precision, the H20 delivers up to 900 TFLOPS, and it also supports FP8 for optimized AI inference.


  • Manufacturer :Enterprise GPUs
  • SpecificationDetails
    GPU ArchitectureNVIDIA Hopper
    GPU Memory96 GB HBM3
    GPU Memory Bandwidth4.0 TB/s
    INT8FPB Tensor Core*
    BF16FP16 Tensor Core®
    TF32 Tensor Core*74 TFLOPS
    FP3244 TFLOPS
    FP641 TFLOPS
    RT CoreN/A
    MIGUp to 7 MIG
    L2 Cache60 MB
    Media Engine7 NVDEC, 7 NVJPEG
    Power400 W
    Form Factor8-way HGX
    InterconnectPCIe Gen5 x16: 128 GB/s, NVLink: 900GB/s
Review this product
Your Rating
Choose File

No reviews available.