Rent GPU
NVIDIA H200

The NVIDIA H200 GPU supercharges generative AI and high-performance computing (HPC) workloads with game-changing performance and memory capabilities. delivering massive throughput improvements for generative AI and large-scale inference workloads.

Memory

141GB
GPU RAM

Bandwidth

4.8TB/s
Memory Bandwidth

Form factor

SXM & NVL
Architecture

Interconnect

900GB/s
NVLink Switch

Accelerated
Workload & Computing

H200 is purpose-built for next-generation AI models, enabling faster inference, reduced latency, and lower total cost for large-scale deployments across data centers and cloud platforms. 1.6x faster for GPT-3 175B Inference, 1.9X Faster for Llama2 70B Inference, 110X Faster for High-Performance Computing

  • Ultra Large Model Inference
  • High Throughput AI Serving
  • Generative AI at Scale

Architecture Comparison

Select the form factor optimized for your specific workload

FeatureH200 SXMH200 NVL
architectureHopperHopper
memory141GB HBM3e141GB HBM3e
memory Bandwidth4.8 TB/s4.8 TB/s
recommended ForBetter Tensor Cores for Advanced Generative AI & LLM InferenceHigh Performance Computing