NVIDIAH200
The NVIDIA H200 is an ideal choice for large-scale AI applications. It uses the NVIDIA Hopper architecture that combines advanced features and capabilities, accelerating AI training and inference on larger models. Starting from $2.49/hr
Available at the most cost-effective pricing
Launch your AI products faster with on-demand GPUs and a global network of data center partners
Virtual machines
The ideal deployment strategy for AI workloads with a H200.
from $2.49/hr
Enquire now- Up to 8 GPUs / virtual machine
- Flexible
- Network attached storage
- Private networks
- Security groups
- Images
Bare metal
Complete control over a physical machine providing absolute performance & control
Pricing available on request
Get pricing- Up to 8 GPUs / host
- No noisy neighbours
- SpectrumX local networking
- 300Gbps external connectivity
- NVMe SSD storage
Looking to scale? Please contact us for enterprise solutions.
Speak with an expertThe NVIDIA H200 is perfect for a wide range of workloads
Deploying AI based workloads on CUDO Compute is easy and cost-effective. Follow our AI related tutorials.
Specifications
Starting from | $2.49/hr |
Architecture | NVIDIA Hopper |
Form factor | SXM |
FP64 tensor core | 34 TFLOPS |
FP32 | 67 TFLOPS |
FP32 tensor core | 67 TFLOPS |
BFLOAT16 tensor core | 1,979 TFLOPS |
FP16 tensor core | 1,979 TFLOPS |
FP8 tensor core | 3,958 TFLOPS |
INT8 tensor core | 3,958 TOPS |
GPU memory | 141GB |
GPU memory bandwidth | 4.8 TB/s |
Decoders | 7x NVDEC 7x NVJPEG |
Max thermal design power (TDP) | Up to 700W |
Multi-instance GPUs (MIG) | Up to 7 MIGs @16.5GB each |
Interconnect | NVLink: 900GB/s PCIe Gen5: 128GB/s |
Use cases
AI inference
AI developers can utilize the NVIDIA H200 to accelerate AI inference workloads, such as image and speech recognition, at lightning speed. The H200 GPU’s powerful Tensor Cores enable it to quickly process large amounts of data, making it perfect for real-time inference applications.
Deep learning
The NVIDIA H200 empowers data scientists and researchers to achieve groundbreaking milestones in deep learning. Its massive memory and processing power guarantee significantly reduced training and deployment times for complex, large-scale models and enables model training on significantly larger datasets.
High-performance computing
From complex scientific simulations to weather forecasting and intricate financial modelling, the H200 empowers diverse organizations to accelerate high-performance computing tasks. Its unmatched memory bandwidth and processing capabilities ensure smooth operation for workloads of any scale, allowing you to achieve unmatched results faster than ever.
Browse alternative GPU solutions for your workloads
Access a wide range of performant NVIDIA and AMD GPUs to accelerate your AI, ML & HPC workloads
An NVIDIA preferred partner for compute
We're proud to be an NVIDIA preferred partner for compute, offering the latest GPUs and high-performance computing solutions.
Also trusted by our other key partners:
Frequently asked questions
Pricing & reservation enquiry
Enquire about access today to test the H200 GPU Cloud, or reserve your H200 Cloud on CUDO Compute for as long as you want it, with unique contracts tailored to suit your needs.
Get started today or speak with an expert...
Available Mon-Fri 9am-5pm UK time