Unprecedented acceleration for the world’s most
demanding AI and machine learning workloads
starting at $2.30 per hour

Availability

Mini Cluster 64 H100 GPUs


Base Cluster 248 H100 GPUs

Pricing

  • Starting at $2.30 per hour

Key Features

  • NVIDIA Quantum-2 3200Gb/s
    InfiniBand Networking
  • Non-Blocking InfiniBand Network
    Design
  • NVIDIA HGX H100 SXM with FP8
    Support

Purpose-built for AI, simulation, and
data analytics

AI, complex simulations, and massive datasets require multiple GPUs with extremely fast interconnections and a fully accelerated software stack. The NVIDIA HGX™ AI supercomputing platform brings together the full power of NVIDIA GPUs, NVLink®, NVIDIA networking, and fully optimized AI and high-performance computing (HPC) software stacks to provide the highest application performance and drive the fastest time to insights.

no form fill or personal details required for access


The world’s most powerful GPU

NVIDIA H200 supercharges generative AI and high-performance computing (HPC) workloads with game-changing performance and memory capabilities. As the first GPU with HBM3e, the H200’s larger and faster memory fuels the acceleration of generative AI and large language models (LLMs) while advancing scientific computing for HPC workloads.

Llama2 70B inference

1.9x faster

GPT-3 175B inference

1.6x faster

High-performance computing

110x faster

NVIDIA GH200
Specifications

NVIDIA H100 SXM NVIDIA H200 SXM1
FP64 34 TFLOPS 34 TFLOPS
FP64 Tensor Core 67 TFLOPS 67 TFLOPS
FP32 67 TFLOPS 67 TFLOPS
TF32 Tensor Core 989 TFLOPS2 989 TFLOPS2
BFLOAT16 Tensor Core 1,979 TFLOPS2 1,979 TFLOPS2
FP16 Tensor Core 1,979 TFLOPS2 1,979 TFLOPS2
FP8 Tensor Core 3,958 TFLOPS2 3,958 TFLOPS2
INT8 Core 3,958 TFLOPS2 3,958 TFLOPS2
GPU Memory 80GB 141GB
GPU Memory Bandwith 3.35TB/s 4.8TB/s
Decoders 7 NVDEC
7JPEG
7 NVDEC
7JPEG
Interconnect NVIDIA NVLink®: 900GB/s
PCIe Gen5: 128GB/s
NVIDIA NVLink®: 900GB/s
PCIe Gen5: 128GB/s
1Preliminary specifications. May be subject to change.
2With sparsity.