AI Engineering Podcast

GPU Clouds, Aggregators, and the New Economics of AI Compute

Jan 27, 2026
Hugo Shi, co-founder and CTO of Saturn Cloud, builds GPU cloud platforms for ML teams. He maps the GPU provider landscape and compares hyperscalers, boutique GPU clouds, bare‑metal and aggregators. He explores orchestration, data gravity, training vs inference splits, accelerator diversity (including AMD progress), supply dynamics, and predictions on consolidation, marketplaces, and reliability for long GPU runs.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Data Gravity Still Rules Migration

  • Data gravity remains the dominant friction for multi-cloud GPU use: copy once or set up continuous caching.
  • Many GPU clouds offer free egress, but most companies keep data in their hyperscaler "home base."
ADVICE

Split Training And Inference Strategically

  • Keep training on hyperscalers when it touches sensitive internal systems, and consider moving inference to GPU clouds to save cost.
  • Balance depends on where your spend is: inference often dominates spend now.
INSIGHT

New Chips Free Up Older Capacity

  • GPU scarcity eased but remains; new generations free up on-demand capacity for previous chips.
  • Increased on-demand supply enables more flexible pricing models like reservations and spot markets on GPU clouds.
Get the Snipd Podcast app to discover more snips from this episode
Get the app