Cloud GPU Pricing Comparison

Compare cloud GPU rental costs across 14+ providers. Find the cheapest H100, A100, L40S, and RTX 4090 instances with on-demand and spot pricing.

Updated Apr 23 60 GPUs ยท 176 prices

Price Insights

๐Ÿš€
Cheapest H100
Vast.ai
$1.38/hr per GPU
โšก
Cheapest A100
Vast.ai
$0.280/hr per GPU
๐Ÿ’ฐ
Best Spot Discount
88% off
RunPod ยท RTX A2000
๐Ÿ“Š
Most GPU Options
RunPod
42 GPU configurations

Cloud GPU Pricing Summary

GPU VRAM Min On-Demand Min Spot Savings Providers
B200 192GB $3.75/hr $3.59/hr 4% off 3 Details โ†’
H200 141GB $2.30/hr $2.29/hr - 7 Details โ†’
H100 SXM 80GB $2.10/hr $1.75/hr 17% off 12 Details โ†’
H100 PCIe 80GB $2.39/hr $1.25/hr 48% off 3 Details โ†’
H100 NVL 94GB $1.38/hr $1.07/hr 23% off 2 Details โ†’
MI300X 192GB $1.99/hr $1.49/hr 25% off 1 Details โ†’
A100 80GB 80GB $1.15/hr $0.950/hr 17% off 9 Details โ†’
A100 40GB 40GB $0.720/hr $1.15/hr -59% off 6 Details โ†’
L40S 48GB $0.860/hr $0.260/hr 70% off 4 Details โ†’
L4 24GB $0.390/hr $0.220/hr 44% off 3 Details โ†’
RTX 4090 24GB $0.235/hr $0.109/hr 54% off 3 Details โ†’
A40 48GB $0.400/hr $0.200/hr 50% off 3 Details โ†’
RTX 4080 16GB $0.500/hr $0.160/hr 68% off 1 Details โ†’
RTX 6000 Ada 48GB $0.750/hr $0.390/hr 48% off 5 Details โ†’
RTX A6000 48GB $0.450/hr $0.330/hr 27% off 6 Details โ†’
RTX 3090 24GB $0.081/hr $0.062/hr 23% off 3 Details โ†’
RTX A4000 16GB $0.060/hr Lowest $0.160/hr -167% off 4 Details โ†’
V100 32GB $0.140/hr $0.992/hr -609% off 5 Details โ†’
T4G 16GB $0.420/hr $0.182/hr 57% off 1 Details โ†’
T4 16GB $0.350/hr $0.140/hr 60% off 3 Details โ†’
A10G 24GB $1.01/hr $0.501/hr 50% off 1 Details โ†’
Gaudi - $1.64/hr $0.638/hr 61% off 1 Details โ†’
A100 40GB 320GB $3.40/hr $1.63/hr 52% off 2 Details โ†’
H200 141GB 1128GB $10.60/hr - - 1 Details โ†’
H100 80GB 640GB $12.29/hr $7.22/hr 41% off 1 Details โ†’
V100 16GB $0.550/hr - - 2 Details โ†’
A100 80GB 80GB $1.39/hr $0.820/hr 41% off 2 Details โ†’
A10 24GB $0.454/hr - - 3 Details โ†’
L40 48GB $0.950/hr $0.500/hr 47% off 3 Details โ†’
RTX A5000 24GB $0.250/hr $0.140/hr 44% off 5 Details โ†’
P100 16GB $1.46/hr $0.555/hr 62% off 1 Details โ†’
P4 8GB $0.600/hr $0.240/hr 60% off 1 Details โ†’
NVIDIA A100 40GB SXM 40GB $1.29/hr - - 1 Details โ†’
NVIDIA GeForce RTX 4070 Ti 12GB $0.500/hr $0.100/hr 80% off 1 Details โ†’
NVIDIA GeForce RTX 3090 Ti 24GB $0.460/hr $0.140/hr 70% off 1 Details โ†’
NVIDIA GeForce RTX 3080 Ti 12GB $0.180/hr $0.090/hr 50% off 1 Details โ†’
NVIDIA GeForce RTX 3080 10GB $0.170/hr $0.090/hr 47% off 1 Details โ†’
B200 SXM6 180GB $4.99/hr - - 1 Details โ†’
A6000 48GB $0.800/hr - - 1 Details โ†’
GH200 96GB $1.49/hr - - 1 Details โ†’
Quadro RTX 6000 24GB $0.500/hr - - 1 Details โ†’
A30 24GB $0.410/hr $0.210/hr 49% off 1 Details โ†’
RTX 3070 8GB $0.130/hr $0.070/hr 46% off 1 Details โ†’
RTX 5080 16GB $0.590/hr $0.300/hr 49% off 1 Details โ†’
RTX 5090 32GB $0.890/hr $0.530/hr 40% off 1 Details โ†’
NVIDIA H200 NVL 143GB $3.39/hr - - 1 Details โ†’
RTX 2000 Ada 16GB $0.240/hr $0.140/hr 42% off 1 Details โ†’
RTX 4000 Ada 20GB $0.260/hr $0.190/hr 27% off 1 Details โ†’
RTX 4000 Ada SFF 20GB $0.440/hr $0.090/hr 80% off 1 Details โ†’
RTX 5000 Ada 32GB $0.830/hr $0.420/hr 49% off 1 Details โ†’
RTX A2000 6GB $0.500/hr $0.060/hr 88% off 1 Details โ†’
RTX A4500 20GB $0.250/hr $0.180/hr 28% off 1 Details โ†’
RTX PRO 6000 MaxQ 96GB $0.500/hr - - 1 Details โ†’
RTX PRO 6000 96GB $1.84/hr $1.19/hr 35% off 1 Details โ†’
RTX PRO 6000 WK 96GB $2.09/hr $1.18/hr 44% off 1 Details โ†’
V100 FHHL 16GB $0.190/hr $0.100/hr 47% off 1 Details โ†’
Tesla V100 16GB $0.190/hr $0.100/hr 47% off 1 Details โ†’
V100 SXM2 16GB $0.230/hr $0.120/hr 48% off 1 Details โ†’
V100 SXM2 32GB 32GB $0.490/hr $0.250/hr 49% off 1 Details โ†’
A100 PCIe 80GB $0.280/hr $0.160/hr 43% off 1 Details โ†’

Provider Types Comparison

Hyperscalers

Enterprise-grade with SLAs, compliance, and global regions. Higher prices but maximum reliability.

GPU Clouds

Specialized for AI/ML workloads. Better GPU availability and competitive pricing.

Marketplaces

Peer-to-peer GPU rental. Lowest prices but variable availability and quality.

Buy vs Rent: Break-Even Analysis

When to Rent Cloud GPUs

  • Usage less than 1,000 hours/year
  • Variable or unpredictable workloads
  • Need different GPU types for different tasks
  • Want to avoid hardware maintenance
  • Need to scale up/down quickly

When to Buy Hardware

  • Usage more than 3,000+ hours/year
  • Consistent, predictable workloads
  • Data privacy/security requirements
  • Long-term projects (2+ years)
  • Have IT infrastructure to manage hardware

๐Ÿ’ก Quick Math

H100 Hardware Cost: ~$30,000
H100 Cloud Cost: ~$3/hr
Break-even Point: ~10,000 hours

That's about 14 months of 24/7 usage. Factor in electricity, cooling, and maintenance for hardware.

Cloud GPU Rental Guide

On-Demand vs Spot

On-Demand:

Pay full price, guaranteed availability. Use for production workloads, demos, and time-sensitive tasks.

Spot/Preemptible:

50-70% cheaper but can be interrupted. Perfect for training jobs with checkpointing, batch processing, and experiments.

Commitment Options

Reserved Instances:

1-3 year commitments for 30-60% savings. Best for predictable, long-term workloads.

Savings Plans:

Commit to spend amount, not specific instances. More flexibility than reserved.

Cost Optimization Tips

  • Use spot instances for training with checkpoints
  • Right-size your instances (don't over-provision)
  • Use auto-scaling for variable workloads
  • Consider GPU clouds for better availability
  • Compare multi-GPU vs single-GPU pricing
  • Check regional pricing differences

Cloud GPU Pricing FAQ

Cloud GPU pricing is typically per-hour based on the GPU type, instance configuration (CPU, RAM, storage), and region. Prices shown here are per-GPU hourly rates in US regions. Multi-GPU instances may have different per-GPU costs due to interconnect and other factors.

On-demand instances are available immediately and run until you stop them - you pay full price for guaranteed availability. Spot (or preemptible) instances can be 50-70% cheaper but may be interrupted when demand is high. Spot is ideal for fault-tolerant workloads like training with checkpoints.

Prices vary frequently, but GPU-specialized clouds like Lambda Labs, CoreWeave, and RunPod typically offer lower H100 prices than hyperscalers (AWS, Azure, GCP). However, hyperscalers may offer better reserved instance discounts for long-term commitments. Check our real-time comparison above.

Yes, spot instances can be reclaimed with short notice (usually 30 seconds to 2 minutes). To handle this, implement checkpointing in your training code to save progress regularly. Most ML frameworks (PyTorch, TensorFlow) support automatic checkpointing.

It depends on your needs. For enterprise with compliance requirements, choose AWS, Azure, or GCP. For best GPU availability and pricing, consider CoreWeave or Lambda Labs. For budget experimentation, try Vast.ai or RunPod. Consider factors like GPU availability, networking speed, and support quality.

Key strategies: 1) Use spot instances for training with checkpoints, 2) Right-size instances (don't over-provision), 3) Use reserved instances for predictable workloads, 4) Compare providers regularly as prices change, 5) Consider GPU clouds over hyperscalers for better rates, 6) Use auto-scaling to avoid idle resources.