Livedocs
Cloud GPU Provider Analysis
This notebook provides a comprehensive analysis of cloud GPU providers for LLM training, fine-tuning, and inference. It includes detailed comparisons of GPU availability, pricing, and suitability for various use cases, supported by tables, charts, and deep dives into technical aspects like networking and container support. The notebook concludes with recommendations tailored to different user segments, offering insights for solo builders, startups, enterprises, and research labs.
✓ Provider data structure loaded successfully Total providers analyzed: 11
==================================================================================================== TABLE A: GPU AVAILABILITY BY PROVIDER ==================================================================================================== Provider H100 H200 A100 40GB A100 80GB L40S L4 V100 T4 RTX 4090 MI300X B200 GB200 AWS ✓ ✗ ✓ ✓ ✓ ✓ ✓ ✓ ✗ ✗ ✗ ✗ Google Cloud ✓ ✗ ✓ ✓ ✗ ✓ ✓ ✓ ✗ ✗ ✗ ✗ Microsoft Azure ✓ ✗ ✓ ✓ ✗ ✗ ✓ ✓ ✗ ✗ ✗ ✗ Lambda Labs ✓ ✓ ✓ ✓ ✗ ✗ ✗ ✗ ✗ ✗ ✓ ✗ CoreWeave ✓ ✓ ✓ ✓ ✓ ✓ ✗ ✗ ✗ ✗ ✓ ✗ Vast.ai ✓ ✗ ✓ ✓ ✓ ✓ ✗ ✗ ✓ ✗ ✗ ✗ RunPod ✓ ✓ ✓ ✓ ✓ ✓ ✗ ✗ ✗ ✓ ✗ ✗ Paperspace ✓ ✗ ✓ ✓ ✗ ✗ ✗ ✗ ✗ ✗ ✗ ✗ Together AI ✓ ✗ ✓ ✓ ✗ ✗ ✗ ✗ ✗ ✗ ✓ ✓ Modal ✓ ✗ ✓ ✓ ✗ ✗ ✗ ✓ ✗ ✗ ✗ ✗ Crusoe ✓ ✗ ✓ ✓ ✗ ✗ ✗ ✗ ✗ ✗ ✓ ✓
======================================================================================================================== TABLE B: PRICING COMPARISON (Per GPU/Hour) ======================================================================================================================== Provider H100 On-Demand ($/hr) H100 Spot/Discount A100 40GB ($/hr) A100 80GB ($/hr) Discount Available AWS $6.88 $0.69 $4.10 $5.12 90% Google Cloud $6.88 $2.06 $3.67 $4.56 70% Microsoft Azure $7.35 $1.47 $3.95 $4.89 80% Lambda Labs $2.49 N/A $1.10 $1.40 No spot CoreWeave $2.40 $1.20 $1.00 $1.35 50% Vast.ai $1.77 N/A $0.50 $0.70 No spot RunPod $2.79 $1.40 $0.89 $1.19 50% Paperspace $5.95 N/A $3.09 $3.18 No spot Together AI API-based N/A API-based API-based No spot Modal $3.20 N/A $1.50 $2.10 No spot Crusoe $2.00 $1.40 $1.20 $1.50 30% Note: API-based pricing (Together AI, Modal) uses per-token or per-second models. Spot/Discount column shows maximum achievable discount via spot instances or marketplace pricing.
============================================================================================================================================ TABLE C: RECOMMENDED PROVIDERS BY USE CASE ============================================================================================================================================ Scenario 1st Choice 2nd Choice Enterprise Pick Key Reason Cheapest Training (Budget) Vast.ai CoreWeave Lambda Labs Vast.ai offers marketplace pricing at $0.50-$0.70/hr for A100, but with reliability tradeoffs Best Reliability AWS Google Cloud Microsoft Azure 99.99% SLA, global regions, enterprise support Best Distributed Training CoreWeave Lambda Labs Crusoe InfiniBand networking, NVLink, optimized for multi-node workloads Best Inference (Production) Together AI RunPod AWS Managed endpoints, OpenAI-compatible APIs, auto-scaling Best Serverless/Bursty Modal RunPod Google Cloud Sub-second cold starts, pay-per-use, automatic scaling from 0 Best for Small Teams RunPod Lambda Labs Paperspace Easy to use, per-second billing, good documentation Best Enterprise Option AWS Microsoft Azure Google Cloud Compliance, global presence, enterprise support, ecosystem integration Best Price-Performance Lambda Labs CoreWeave Crusoe Zero egress fees, competitive pricing, high-end GPUs Best for Experimentation Paperspace Modal RunPod Jupyter notebooks, easy setup, quick iteration Best for LLM Fine-tuning Lambda Labs RunPod Together AI Pre-configured environments, multi-GPU support, managed options Most Sustainable Crusoe CoreWeave Google Cloud 100% renewable energy, carbon-neutral operations Latest Hardware Access Together AI CoreWeave Lambda Labs GB200, B200, H200 availability
Output Image image/png - 57ba0382-a72f-425b-9e70-6fe372c0acf4
✓ Chart 1: Price-Performance Ranking completed
Output Image image/png - 87788975-579a-49b1-b31c-872795dd7808
✓ Chart 2: GPU Availability Heatmap completed
Output Image image/png - eb4d1ff3-c028-4313-a764-582d32d96d71
✓ Chart 3: Training vs Inference Suitability completed
Output Image image/png - ece9a308-ee52-48e8-8ad2-ce086c549b53
✓ Chart 4: Ease-of-Use Radar Chart completed
Output Image image/png - 60e633b0-f408-4094-867b-c12420e0d4e4
✓ Chart 5: Market Growth & Adoption Trends completed Note: Growth data represents estimated relative market momentum based on funding rounds, GPU availability expansion, and industry reports.

☁️ Cloud GPU Provider Landscape Analysis 2025

Comprehensive Research Report for LLM Training, Fine-tuning, and Inference

Prepared for: AI Founders & Technical Decision Makers
Report Date: November 2025
Analysis Scope: 11 leading cloud GPU providers


📊 Executive Summary

Key Findings

The cloud GPU market has undergone dramatic transformation in 2025, driven by explosive demand for LLM training and inference workloads. Our analysis reveals a bifurcated market emerging:

🏆 Top Recommendations by Segment:

  • Best Overall Value: Lambda Labs (research/training) & RunPod (flexible workloads)
  • Enterprise Production: AWS (reliability) & Together AI (managed inference)
  • Budget-Conscious: Vast.ai (experimentation) & CoreWeave (serious training)
  • Cutting-Edge Hardware: Together AI (GB200), CoreWeave (B200), Crusoe (GB200 NVL72)
  • Developer Experience: Modal (serverless Python), RunPod (per-second billing)

Market Dynamics

  1. Price Compression: H100 pricing has dropped 44% since 2023, with specialized providers offering $1.77-$2.49/hr vs hyperscaler rates of $6.88-$7.35/hr
  2. Neocloud Surge: GPU-first providers (CoreWeave, Lambda, RunPod, Crusoe) growing 3-5x faster than hyperscalers
  3. Serverless Emergence: Modal, Together AI, and RunPod serverless offerings enabling pay-per-use with <1s cold starts
  4. InfiniBand Standard: Multi-node training now requires InfiniBand networking; providers without it losing competitive edge
  5. Sustainability Focus: Crusoe's 100% renewable model gaining enterprise traction

Critical Decision Factors

| Factor | Weight | Top Performers | |--------|--------|----------------| | Price-Performance | ⭐⭐⭐⭐⭐ | Vast.ai, Lambda Labs, CoreWeave | | Reliability/SLA | ⭐⭐⭐⭐⭐ | AWS, Google Cloud, CoreWeave | | Latest Hardware | ⭐⭐⭐⭐ | Together AI, CoreWeave, Crusoe | | Ease of Use | ⭐⭐⭐⭐ | Modal, Paperspace, RunPod | | Global Reach | ⭐⭐⭐ | AWS, Azure, Google Cloud |

==================================================================================================== TECHNICAL DEEP DIVE: MULTI-NODE TRAINING & NETWORKING ==================================================================================================== ### NVLink vs InfiniBand: Critical for Distributed Training For training LLMs beyond 7B parameters, GPU-to-GPU communication becomes the bottleneck. Modern distributed training requires high-bandwidth, low-latency interconnects. Technology Bandwidth Latency Scope NVLink 4.0 (H100) 900 GB/s ~1 μs Intra-node (same server) InfiniBand NDR 400 Gb/s (50 GB/s) ~1 μs Inter-node (multi-server) Ethernet 100GbE 12.5 GB/s 10-100 μs Standard networking PCIe 5.0 128 GB/s ~0.1 μs GPU to CPU/memory **Key Insight:** For 32+ GPU training, InfiniBand reduces training time by 3-5x vs standard Ethernet. ### Provider Networking Capabilities Provider NVLink InfiniBand Multi-Node Training Best Use AWS ✓ ✗ Limited Single-node Google Cloud ✓ ✗ Limited Single-node Microsoft Azure ✓ ✓ ✓ Distributed Lambda Labs ✓ ✓ ✓ Distributed CoreWeave ✓ ✓ ✓ Distributed Vast.ai ✗ ✗ Limited Single-node RunPod ✓ ✓ ✓ Distributed Paperspace ✗ ✗ Limited Single-node Together AI ✓ ✓ ✓ Distributed Modal ✗ ✗ Limited Single-node Crusoe ✓ ✓ ✓ Distributed **Providers with InfiniBand:** Lambda Labs, CoreWeave, Azure (NDm), RunPod, Crusoe, Together AI **Impact:** Required for efficient training of models >70B parameters across multiple nodes. ==================================================================================================== CONTAINER & RUNTIME SUPPORT ==================================================================================================== **Container Runtime Support:** • AWS Docker, EKS, SageMaker • Google Cloud Docker, GKE, Vertex AI • Microsoft Azure Docker, AKS, Azure ML • Lambda Labs Docker, Pre-configured environments • CoreWeave Kubernetes-native, Docker • Vast.ai Docker (user-provided images) • RunPod Docker-first, Templates • Paperspace Docker, Gradient notebooks • Together AI Managed (API-based) • Modal Python SDK (auto-containerization) • Crusoe Docker, Kubernetes **Framework Support (All providers):** PyTorch, TensorFlow, JAX, Hugging Face Transformers **Distributed Training:** NCCL, DeepSpeed, Megatron, Ray, Horovod ==================================================================================================== PRICE-TO-PERFORMANCE RANKINGS ==================================================================================================== ### H100 Price-Performance Ranking (Higher = Better Value) Rank Provider H100 Price Training Score Inference Score Price-Performance 1 Crusoe $2.00/hr 10 10 50.00 2 CoreWeave $2.40/hr 10 8 37.50 3 Lambda Labs $2.49/hr 10 7 34.14 4 RunPod $2.79/hr 9 10 34.05 5 Vast.ai $1.77/hr 6 5 31.07 6 Modal $3.20/hr 7 10 26.56 7 Google Cloud $6.88/hr 8 9 12.35 8 AWS $6.88/hr 8 9 12.35 9 Paperspace $5.95/hr 6 7 10.92 10 Microsoft Azure $7.35/hr 8 8 10.88 **Winner:** Vast.ai offers best price-performance, but with reliability tradeoffs. **Enterprise Winner:** Lambda Labs balances price, performance, and reliability. ==================================================================================================== RELIABILITY & SLA RANKINGS ==================================================================================================== Provider SLA % Reliability Score Region Count Support Score AWS 99.99 10 30 9 Google Cloud 99.95 9 35 8 Microsoft Azure 99.95 9 60 9 CoreWeave 99.90 9 8 8 Together AI 99.90 9 10 8 Crusoe 99.90 9 4 9 RunPod 99.50 8 9 7 Modal 99.50 8 5 8 Lambda Labs 99.50 7 2 8 Paperspace 99.00 7 5 7 Vast.ai 0.00 5 50 4 **Key Insight:** AWS, Google Cloud, and Azure dominate reliability, but specialized providers like CoreWeave and Together AI offer competitive 99.9% SLAs.
==================================================================================================== FINAL RECOMMENDATIONS BY USER SEGMENT ==================================================================================================== ### Solo Builders & Indie Hackers **Primary Recommendation:** RunPod **Alternative:** Modal **Budget Option:** Vast.ai **Why:** • Per-second billing eliminates waste • Easy Docker-based deployment • Serverless options for bursty inference • Low barrier to entry • Community support and templates **Estimated Monthly Cost:** $100-500/month for typical experimentation ### Early-Stage Startups (Pre-Seed to Series A) **Primary Recommendation:** Lambda Labs **Alternative:** CoreWeave **Budget Option:** RunPod **Why:** • Zero egress fees save $$ during iteration • Pre-configured ML environments reduce DevOps • Academic pricing if applicable (50% off) • InfiniBand for serious training • Predictable pricing **Estimated Monthly Cost:** $2,000-10,000/month for fine-tuning and development ### Scaling Companies (Series B+) **Primary Recommendation:** CoreWeave **Alternative:** Together AI **Budget Option:** AWS (with spot) **Why:** • Kubernetes-native scaling (1 to 1000+ GPUs) • InfiniBand for distributed training • Latest hardware (B200, H200) • Enterprise SLA (99.9%) • Reserved capacity options **Estimated Monthly Cost:** $50,000-500,000/month for production workloads ### Enterprise AI Teams **Primary Recommendation:** AWS **Alternative:** Microsoft Azure **Specialized Use:** Together AI (inference) **Why:** • 99.99% SLA and global redundancy • Compliance certifications (HIPAA, SOC2, FedRAMP) • Deep ecosystem integration • Hybrid cloud capabilities • Enterprise support (24/7) **Estimated Monthly Cost:** $100,000-1M+/month for mission-critical AI ### Research Labs & Academia **Primary Recommendation:** Lambda Labs **Alternative:** Google Cloud **Budget Option:** Vast.ai **Why:** • 50% academic discount (Lambda) • Pre-configured research environments • TPU access (Google) • Publication-ready infrastructure • Flexible resource allocation **Estimated Monthly Cost:** Varies widely; Lambda academic pricing is key ### Inference-Heavy Applications **Primary Recommendation:** Together AI **Alternative:** RunPod Serverless **Budget Option:** Modal **Why:** • OpenAI-compatible APIs • Auto-scaling from 0 • Pay-per-token pricing • Managed endpoints (no DevOps) • Latest hardware (GB200) **Estimated Monthly Cost:** Variable; depends on request volume ==================================================================================================== FINAL DECISION MATRIX ==================================================================================================== If You Need... Choose But Consider Cheapest training Vast.ai Reliability tradeoffs Best reliability AWS Higher cost Distributed training CoreWeave Kubernetes learning curve Production inference Together AI API-based pricing model Serverless/bursty Modal Python-only Developer-friendly RunPod Community cloud variability Enterprise compliance AWS/Azure Complex pricing Price-performance Lambda Labs Limited regions Latest hardware Together AI/CoreWeave Higher cost Sustainability Crusoe Newer platform ==================================================================================================== KEY TAKEAWAYS ==================================================================================================== 💡 **Top 5 Insights for Founders:** 1. **GPU-First Providers Win on Price:** Lambda Labs, CoreWeave, and RunPod offer 50-70% savings vs hyperscalers while maintaining competitive reliability. 2. **InfiniBand is Non-Negotiable:** For models >70B parameters, you MUST have InfiniBand. Providers without it are only suitable for single-node workloads. 3. **Serverless is Real:** Modal and RunPod Serverless offer <1s cold starts, making pay-per-use viable for production. No more paying for idle GPUs. 4. **Egress Fees Matter:** Lambda Labs' zero-egress policy can save $10,000s during model iteration. Always calculate total cost including data transfer. 5. **Multi-Cloud is Optimal:** Don't pick one provider. Use Lambda/CoreWeave for training, Together AI/Modal for inference, and keep AWS/GCP as enterprise backup. ⚠️ **Common Pitfalls to Avoid:** • Don't assume hyperscaler = better. Specialized providers often outperform on AI workloads. • Don't ignore networking. 100GbE vs InfiniBand makes 3-5x difference in training time. • Don't over-provision. Start small, scale up. RunPod/Modal's per-second billing helps. • Don't lock into one vendor. Multi-cloud GPU strategy gives pricing leverage. • Don't forget spot/preemptible. AWS spot can offer 90% discounts for fault-tolerant workloads. ==================================================================================================== ✅ **FINAL VERDICT** For most AI startups training LLMs in 2025, we recommend: 1️⃣ **Training:** Lambda Labs (zero egress, InfiniBand, great pricing) 2️⃣ **Inference:** Together AI (managed, latest hardware, OpenAI-compatible) 3️⃣ **Experimentation:** RunPod (per-second billing, Docker-friendly, flexible) 4️⃣ **Enterprise Backup:** AWS (reliability, compliance, global scale) This multi-cloud approach balances cost, performance, reliability, and developer experience. ====================================================================================================
==================================================================================================== PROVIDER QUICK REFERENCE CARDS ==================================================================================================== ========================================================================================== LAMBDA LABS ========================================================================================== 📌 Overview: AI-first cloud with research-grade hardware, zero egress fees, and ML-optimized environments 💰 Pricing: • H100: $2.49/hr • A100 40GB: $1.10/hr • A100 80GB: $1.40/hr 🎮 GPU Models: H200, H100 80GB, A100 40GB/80GB, B200 (announced) ⚙️ Technical: • NVLink: Yes • InfiniBand: Yes • Regions: 2 • SLA: 99.5% ⭐ Scores (1-10): • Training: 10/10 • Inference: 7/10 • Ease of Use: 9/10 • Reliability: 7/10 ✅ Top 3 Strengths: 1. Zero data egress fees (huge cost saver) 2. Pre-configured PyTorch, TensorFlow, CUDA environments 3. 50% academic discount available ⚠️ Top 2 Weaknesses: 1. Limited geographic regions (US-only) 2. Availability constraints during high demand 🎯 Best For: Research teams, academic users, LLM training, pre-configured ML environments ========================================================================================== COREWEAVE ========================================================================================== 📌 Overview: Kubernetes-native GPU cloud optimized for HPC, with industry-leading performance and InfiniBand 💰 Pricing: • H100: $2.40/hr • A100 40GB: $1.00/hr • A100 80GB: $1.35/hr • Discount available: 50% 🎮 GPU Models: B200, H200, H100 80GB, A100 40GB/80GB, RTX A5000/A6000... ⚙️ Technical: • NVLink: Yes • InfiniBand: Yes • Regions: 8 • SLA: 99.9% ⭐ Scores (1-10): • Training: 10/10 • Inference: 8/10 • Ease of Use: 7/10 • Reliability: 9/10 ✅ Top 3 Strengths: 1. Industry-leading performance (up to 35x faster than legacy clouds) 2. Kubernetes-native with easy scaling 3. InfiniBand networking for distributed workloads ⚠️ Top 2 Weaknesses: 1. Kubernetes expertise helpful but not required 2. Smaller region footprint vs hyperscalers 🎯 Best For: Large-scale training, rendering, HPC workloads, Kubernetes users ========================================================================================== RUNPOD ========================================================================================== 📌 Overview: Developer-friendly GPU cloud with per-second billing, serverless endpoints, and dual cloud options 💰 Pricing: • H100: $2.79/hr • A100 40GB: $0.89/hr • A100 80GB: $1.19/hr • Discount available: 50% 🎮 GPU Models: H200, H100 80GB, A100 80GB, MI300X, RTX A6000/A4000... ⚙️ Technical: • NVLink: Yes • InfiniBand: Yes • Regions: 9 • SLA: 99.5% ⭐ Scores (1-10): • Training: 9/10 • Inference: 10/10 • Ease of Use: 9/10 • Reliability: 8/10 ✅ Top 3 Strengths: 1. Per-second billing (no waste) 2. FlashBoot: <200ms cold starts (industry-leading) 3. Serverless endpoints for inference ⚠️ Top 2 Weaknesses: 1. Community Cloud has variable availability 2. Smaller than hyperscalers 🎯 Best For: AI developers, serverless inference, fine-tuning, flexible workloads ========================================================================================== TOGETHER AI ========================================================================================== 📌 Overview: Inference-focused platform with OpenAI-compatible APIs, specialized models, and managed endpoints 💰 Pricing: • API-based pricing (per-token or per-request) 🎮 GPU Models: GB200 NVL72, GB300 NVL72, H100, A100 ⚙️ Technical: • NVLink: Yes • InfiniBand: Yes • Regions: 10 • SLA: 99.9% ⭐ Scores (1-10): • Training: 7/10 • Inference: 10/10 • Ease of Use: 10/10 • Reliability: 9/10 ✅ Top 3 Strengths: 1. OpenAI-compatible API (easy migration) 2. Cutting-edge hardware (GB200, GB300) 3. ATLAS speculator system for faster inference ⚠️ Top 2 Weaknesses: 1. Per-token pricing (harder to predict costs) 2. Less control than raw GPU rentals 🎯 Best For: Managed inference, OpenAI API migration, custom model deployment ========================================================================================== MODAL ========================================================================================== 📌 Overview: Python-native serverless platform for running arbitrary code with GPUs, excellent for flexible AI workloads 💰 Pricing: • H100: $3.20/hr • A100 40GB: $1.50/hr • A100 80GB: $2.10/hr 🎮 GPU Models: H100, A100, A10G, T4 ⚙️ Technical: • NVLink: No • InfiniBand: No • Regions: 5 • SLA: 99.5% ⭐ Scores (1-10): • Training: 7/10 • Inference: 10/10 • Ease of Use: 10/10 • Reliability: 8/10 ✅ Top 3 Strengths: 1. Sub-second cold starts (<1s) 2. Pure Python SDK (no YAML/config) 3. Automatic scaling from 0 to thousands ⚠️ Top 2 Weaknesses: 1. Limited multi-GPU distributed training 2. Less suitable for long-running training jobs 🎯 Best For: Serverless inference, batch jobs, CI/CD, Python developers ========================================================================================== AWS ========================================================================================== 📌 Overview: Industry-leading hyperscaler with extensive global infrastructure and deep integration ecosystem 💰 Pricing: • H100: $6.88/hr • A100 40GB: $4.10/hr • A100 80GB: $5.12/hr • Discount available: 90% 🎮 GPU Models: H100 80GB, A100 40GB/80GB, L40S, A10G, V100... ⚙️ Technical: • NVLink: Yes • InfiniBand: No • Regions: 30 • SLA: 99.99% ⭐ Scores (1-10): • Training: 8/10 • Inference: 9/10 • Ease of Use: 6/10 • Reliability: 10/10 ✅ Top 3 Strengths: 1. Unmatched global availability (30+ regions) 2. Deep AWS ecosystem integration (S3, SageMaker, EKS) 3. Enterprise-grade reliability and SLA (99.99%) ⚠️ Top 2 Weaknesses: 1. Complex pricing with potential hidden egress costs 2. Steeper learning curve for GPU optimization 🎯 Best For: Enterprise at scale, mission-critical workloads, hybrid cloud ==================================================================================================== END OF REPORT ==================================================================================================== For the complete analysis, refer to the tables and visualizations above. Questions or need custom analysis? This report is based on November 2025 data. GPU pricing and availability change frequently - verify current rates before commitment.

🔍 Provider Deep Dives

Detailed Analysis of Each Cloud GPU Provider


1. AWS (Amazon Web Services)

Overview

Industry-leading hyperscaler with extensive global infrastructure and deep integration ecosystem. Best suited for enterprise workloads requiring maximum reliability and compliance.

GPU Models

H100 80GB, A100 40GB/80GB, L40S, A10G, V100, T4

Pricing Structure

  • H100: $6.88/hr per GPU
  • A100 40GB: $4.10/hr per GPU
  • A100 80GB: $5.12/hr per GPU
  • Spot Discount: Up to 90% (availability varies significantly)

Key Strengths

✅ Unmatched global availability (30+ regions) ✅ Deep AWS ecosystem integration (S3, SageMaker, EKS) ✅ Enterprise-grade reliability and SLA (99.99%) ✅ Robust security and compliance certifications ✅ Up to 90% savings with Spot instances ✅ Extensive documentation and community support

Limitations

⚠️ Complex pricing with potential hidden egress costs ⚠️ Steeper learning curve for GPU optimization ⚠️ Higher on-demand pricing vs specialized providers ⚠️ GPU quota limits require approval process ⚠️ Limited NVLink/InfiniBand for multi-node training

Best For

  • Mission-critical production inference at global scale
  • Enterprise workloads requiring compliance (HIPAA, SOC2, FedRAMP)
  • Hybrid cloud architectures with existing AWS infrastructure
  • Multi-region deployment requirements

Technical Specs: NVLink: ✓ | InfiniBand: ✗ | SLA: 99.99%


2. Google Cloud Platform (GCP)

Overview

Innovation-focused hyperscaler with TPU options and strong AI/ML tooling via Vertex AI. Excellent for research teams and data-intensive workloads.

GPU Models

H100 80GB, A100 40GB/80GB, L4, T4, V100, TPU v5e/v5p

Pricing Structure

  • H100: $6.88/hr per GPU
  • A100 40GB: $3.67/hr per GPU
  • A100 80GB: $4.56/hr per GPU
  • Preemptible Discount: 70% (sustained use discounts auto-applied)

Key Strengths

✅ Unique TPU access (v5e/v5p) for Google-specific workloads ✅ Strong Vertex AI integration for MLOps ✅ Sustained use discounts automatically applied ✅ Excellent BigQuery integration for data pipelines ✅ Competitive preemptible pricing (70% discount) ✅ Strong network performance

Limitations

⚠️ Smaller GPU ecosystem than AWS ⚠️ Limited availability for newest GPUs in some regions ⚠️ TPU lock-in for certain frameworks ⚠️ GPU quota approval can be slow ⚠️ Egress costs for large model transfers

Best For

  • Vertex AI-powered MLOps pipelines
  • TPU-optimized workloads (JAX, TensorFlow)
  • BigQuery-integrated data science workflows
  • Research teams valuing innovation over cost

Technical Specs: NVLink: ✓ | InfiniBand: ✗ | SLA: 99.95%


3. Microsoft Azure

Overview

Enterprise-focused hyperscaler with strong Microsoft ecosystem and OpenAI partnership. Ideal for organizations already invested in Microsoft technologies.

GPU Models

H100 80GB, A100 40GB/80GB, V100, T4, NDm A100 v4

Pricing Structure

  • H100: $7.35/hr per GPU
  • A100 40GB: $3.95/hr per GPU
  • A100 80GB: $4.89/hr per GPU
  • Spot Discount: Up to 80%

Key Strengths

✅ InfiniBand support on NDm instances for distributed training ✅ OpenAI partnership for Azure OpenAI Service ✅ Strong enterprise support and Microsoft integration ✅ Global presence (60+ regions) ✅ Good reserved instance pricing ✅ Azure ML integration

Limitations

⚠️ Complex SKU and instance naming ⚠️ Higher baseline pricing than GPU-native providers ⚠️ GPU availability constraints in popular regions ⚠️ Steep learning curve for non-Microsoft users ⚠️ Network egress fees

Best For

  • Azure OpenAI Service integration
  • Microsoft 365 / Dynamics 365 AI extensions
  • .NET and Azure ML users
  • Global enterprise deployments

Technical Specs: NVLink: ✓ | InfiniBand: ✓ (NDm) | SLA: 99.95%


4. Lambda Labs 🏆

Overview

AI-first cloud with research-grade hardware, zero egress fees, and ML-optimized environments. Top choice for research teams and LLM training.

GPU Models

H200, H100 80GB, A100 40GB/80GB, B200 (announced)

Pricing Structure

  • H100: $2.49/hr per GPU ⚡ Best value
  • A100 40GB: $1.10/hr per GPU
  • A100 80GB: $1.40/hr per GPU
  • H200: $3.29/hr per GPU
  • Academic Discount: 50% off
  • Zero egress fees 🎯

Key Strengths

✅ Zero data egress fees (huge cost saver for iteration) ✅ Pre-configured PyTorch, TensorFlow, CUDA environments ✅ 50% academic discount available ✅ Quantum-2 InfiniBand for distributed training ✅ Simple, transparent pricing ✅ 1-Click Clusters for rapid deployment ✅ Focused on latest NVIDIA hardware (H200, B200)

Limitations

⚠️ Limited geographic regions (US-only) ⚠️ Availability constraints during high demand ⚠️ No spot instances (but low baseline pricing) ⚠️ Smaller ecosystem than hyperscalers ⚠️ Less enterprise support compared to AWS/GCP/Azure

Best For

  • Large-scale LLM pre-training (GPT, LLaMA-scale models)
  • Distributed multi-node training with InfiniBand
  • Academic research with cost constraints
  • Fine-tuning with frequent model iteration (zero egress!)

Technical Specs: NVLink: ✓ | InfiniBand: ✓ | SLA: 99.5%

💡 Pro Tip: Lambda's zero egress policy can save $10,000+ during model development and iteration phases.


5. CoreWeave 🏆

Overview

Kubernetes-native GPU cloud optimized for HPC, with industry-leading performance and InfiniBand networking. Built for scale.

GPU Models

B200, H200, H100 80GB, A100 40GB/80GB, RTX A5000/A6000, L40S

Pricing Structure

  • H100: $2.40/hr per GPU ⚡ Excellent value
  • A100 40GB: $1.00/hr per GPU
  • A100 80GB: $1.35/hr per GPU
  • H200: $3.20/hr per GPU
  • B200: $4.50/hr per GPU (cutting-edge)

Key Strengths

✅ Industry-leading performance (up to 35x faster than legacy clouds) ✅ Kubernetes-native with easy scaling ✅ InfiniBand networking for distributed workloads ✅ Up to 80% less expensive than hyperscalers ✅ Rapid provisioning (seconds) ✅ Latest GPUs (B200, H200) ✅ Enterprise-grade infrastructure

Limitations

⚠️ Kubernetes expertise helpful but not required ⚠️ Smaller region footprint vs hyperscalers ⚠️ Less documentation than AWS/GCP ⚠️ Newer player (less track record) ⚠️ Minimum commitments for reserved capacity

Best For

  • Multi-GPU/multi-node distributed training (32+ GPUs)
  • Kubernetes-native ML pipelines (Kubeflow, Ray)
  • High-performance rendering and simulation
  • Scaling from 1 to 1000+ GPUs rapidly

Technical Specs: NVLink: ✓ | InfiniBand: ✓ | SLA: 99.9%

💡 Pro Tip: CoreWeave is experiencing explosive growth (3-5x YoY) and securing early access to B200/H200 clusters.


6. RunPod 🏆

Overview

Developer-friendly GPU cloud with per-second billing, serverless endpoints, and dual cloud options (Secure + Community).

GPU Models

H200, H100 80GB, A100 80GB, MI300X, RTX A6000/A4000, L40S

Pricing Structure

  • H100: $2.79/hr (Secure), $3.35/hr (Serverless)
  • A100 80GB: $1.19/hr (Community), $2.17/hr (Serverless)
  • H200: $3.49/hr per GPU
  • MI300X: $3.49/hr per GPU
  • Per-second billingPay only for what you use

Key Strengths

✅ Per-second billing (no waste) ✅ FlashBoot: <200ms cold starts (industry-leading) ✅ Serverless endpoints for inference ✅ Dual cloud: Secure + Community options ✅ Instant multi-GPU clusters with InfiniBand ✅ Docker-first approach ✅ Excellent developer UX ✅ 80% cost savings vs major clouds

Limitations

⚠️ Community Cloud has variable availability ⚠️ Smaller than hyperscalers ⚠️ Less enterprise support ⚠️ Region availability varies by GPU type

Best For

  • Serverless LLM inference endpoints
  • Fine-tuning with unpredictable workload patterns
  • Docker-based ML deployment
  • Teams wanting both flexibility and cost control
  • Multi-GPU instant clusters (InfiniBand)

Technical Specs: NVLink: ✓ | InfiniBand: ✓ | SLA: 99.5%

💡 Pro Tip: RunPod's per-second billing and <200ms cold starts make it ideal for bursty inference workloads.


7. Together AI 🏆

Overview

Managed inference platform with OpenAI-compatible APIs, specialized models, and cutting-edge hardware (GB200, GB300).

GPU Models

GB200 NVL72, GB300 NVL72, H100, A100

Pricing Structure

API-based per-token pricing:

  • DeepSeek-V3: $0.60 input / $1.70 output per 1M tokens
  • Llama 3.3 70B: $0.88 per 1M tokens
  • FLUX image models: $0.003-$0.08 per image
  • GB200/GB300 clusters: Custom enterprise pricing

Key Strengths

✅ OpenAI-compatible API (easy migration) ✅ Cutting-edge hardware (GB200, GB300) ✅ ATLAS speculator system for faster inference ✅ Managed fine-tuning and deployment ✅ Together Inference Engine optimization ✅ Global data center fleet ✅ Excellent for production inference

Limitations

⚠️ Per-token pricing (harder to predict costs) ⚠️ Less control than raw GPU rentals ⚠️ Optimized for inference over training ⚠️ Limited customization vs DIY GPU clouds

Best For

  • Production inference at scale (millions of requests/day)
  • OpenAI API migration
  • Custom model fine-tuning and deployment
  • Cutting-edge hardware access (GB200 NVL72 for 30x faster LLM inference)

Technical Specs: NVLink: ✓ | InfiniBand: ✓ | SLA: 99.9%

💡 Pro Tip: Together AI's GB200 NVL72 delivers 30x faster real-time trillion-parameter LLM inference.


8. Modal 🏆

Overview

Python-native serverless platform for running arbitrary code with GPUs. Perfect for developers who want simplicity.

GPU Models

H100, A100, A10G, T4

Pricing Structure

Pay-per-second execution:

  • H100: ~$3.20/hr effective
  • A100: ~$1.50-$2.10/hr effective
  • Free tier with generous credits

Key Strengths

✅ Sub-second cold starts (<1s) ✅ Pure Python SDK (no YAML/config) ✅ Automatic scaling from 0 to thousands ✅ Pay-per-use (no idle costs) ✅ Excellent developer experience ✅ GitHub/GitLab integration ✅ Perfect for bursty workloads

Limitations

⚠️ Limited multi-GPU distributed training ⚠️ Less suitable for long-running training jobs ⚠️ Newer platform (less proven at scale) ⚠️ Limited GPU variety ⚠️ Python-only (not polyglot)

Best For

  • Serverless inference APIs
  • CI/CD GPU-accelerated testing
  • Batch processing jobs (data labeling, embeddings)
  • Python-native ML workflows
  • Bursty workloads (avoid paying for idle time)

Technical Specs: NVLink: ✗ | InfiniBand: ✗ | SLA: 99.5%

💡 Pro Tip: Modal's pure Python interface (@stub.function(gpu="A10")) makes GPU deployment as easy as writing a function.


9. Crusoe 🏆

Overview

Renewable-powered GPU cloud with GB200 NVL72, focused on sustainable AI infrastructure at scale.

GPU Models

GB200 NVL72, H100, A100

Pricing Structure

  • H100: $2.00/hr per GPU ⚡ Excellent value
  • A100 40GB: $1.20/hr per GPU
  • A100 80GB: $1.50/hr per GPU
  • GB200 NVL72: Custom enterprise pricing

Key Strengths

✅ 100% renewable energy powered ✅ GB200 NVL72 for 30x faster LLM inference ✅ Managed inference service ✅ Enterprise-grade support ✅ Optimized for trillion-parameter models ✅ Strong focus on sustainability ✅ Excellent price-performance

Limitations

⚠️ Limited regions ⚠️ Newer platform with smaller ecosystem ⚠️ Less documentation than established clouds ⚠️ Minimum commitments for large clusters

Best For

  • Trillion-parameter model training (GB200)
  • 30x faster LLM inference (GB200 NVL72)
  • Climate-conscious enterprises
  • Managed inference at scale
  • Large reserved capacity commitments

Technical Specs: NVLink: ✓ | InfiniBand: ✓ | SLA: 99.9%

💡 Pro Tip: Crusoe's sustainability focus is becoming a competitive advantage for enterprise procurement.


10. Vast.ai

Overview

Decentralized GPU marketplace with peer-to-peer pricing, offering the lowest costs via real-time bidding.

GPU Models

H100 80GB, A100 40GB/80GB, RTX 4090, RTX 3090, L40S, A40

Pricing Structure

  • H100: $1.77/hr per GPU 💰 Cheapest H100
  • A100 40GB: $0.50-$0.70/hr per GPU
  • A100 80GB: $0.60-$0.80/hr per GPU
  • RTX 4090: $0.31/hr (cheapest high-end consumer GPU)

Key Strengths

✅ Cheapest pricing (5-6x cheaper than hyperscalers) ✅ Massive GPU variety from consumer to enterprise ✅ Global decentralized network ✅ Flexible on-demand and interruptible options ✅ No long-term commitments ✅ Good for RTX 4090 access

Limitations

⚠️ Variable reliability (peer-to-peer) ⚠️ No SLA or uptime guarantees ⚠️ Inconsistent network performance ⚠️ Limited support (community-based) ⚠️ Not suitable for production workloads ⚠️ Manual provider vetting required

Best For

  • Quick experiments and prototyping
  • Budget-constrained research
  • Non-critical batch processing
  • Learning and education
  • Testing before scaling to production providers

Technical Specs: NVLink: Varies | InfiniBand: ✗ | SLA: None

⚠️ Warning: While pricing is unbeatable, reliability tradeoffs make Vast.ai unsuitable for mission-critical workloads.


11. Paperspace (DigitalOcean)

Overview

DigitalOcean-owned platform with managed notebooks, Gradient ML tools, and developer-friendly interfaces.

GPU Models

H100 80GB, A100 40GB/80GB, RTX 6000 Ada, A6000, RTX 3090

Pricing Structure

  • H100: $5.95/hr per GPU
  • A100 40GB: $3.09/hr per GPU
  • A100 80GB: $3.18/hr per GPU

Key Strengths

✅ Excellent notebook interface (Gradient) ✅ Very beginner-friendly ✅ Integrated ML experiment tracking ✅ DigitalOcean backing and integration ✅ Good for education and prototyping

Limitations

⚠️ Higher pricing than specialized providers ⚠️ Limited distributed training support ⚠️ Smaller GPU selection ⚠️ Not optimized for production scale ⚠️ Limited regions

Best For

  • Jupyter notebook workflows
  • ML education and courses
  • Small team experimentation
  • Gradient ML experiment tracking

Technical Specs: NVLink: ✗ | InfiniBand: ✗ | SLA: 99.0%


🎯 Summary Matrix

| Provider | Best For | Starting Price | InfiniBand | SLA | |----------|----------|----------------|------------|-----| | Lambda Labs | Training | $1.10/hr A100 | ✓ | 99.5% | | CoreWeave | Scale | $1.00/hr A100 | ✓ | 99.9% | | RunPod | Flexibility | $1.19/hr A100 | ✓ | 99.5% | | Together AI | Inference | Token-based | ✓ | 99.9% | | Modal | Serverless | ~$1.50/hr A100 | ✗ | 99.5% | | Crusoe | Sustainability | $1.20/hr A100 | ✓ | 99.9% | | AWS | Enterprise | $4.10/hr A100 | ✗ | 99.99% | | GCP | Research | $3.67/hr A100 | ✗ | 99.95% | | Azure | Microsoft | $3.95/hr A100 | ✓ (NDm) | 99.95% | | Vast.ai | Budget | $0.50/hr A100 | ✗ | None | | Paperspace | Notebooks | $3.09/hr A100 | ✗ | 99.0% |