
AI Engineering Podcast GPU Clouds, Aggregators, and the New Economics of AI Compute
Jan 27, 2026
Hugo Shi, co-founder and CTO of Saturn Cloud, builds GPU cloud platforms for ML teams. He maps the GPU provider landscape and compares hyperscalers, boutique GPU clouds, bare‑metal and aggregators. He explores orchestration, data gravity, training vs inference splits, accelerator diversity (including AMD progress), supply dynamics, and predictions on consolidation, marketplaces, and reliability for long GPU runs.
AI Snips
Chapters
Transcript
Episode notes
Data Gravity Still Rules Migration
- Data gravity remains the dominant friction for multi-cloud GPU use: copy once or set up continuous caching.
- Many GPU clouds offer free egress, but most companies keep data in their hyperscaler "home base."
Split Training And Inference Strategically
- Keep training on hyperscalers when it touches sensitive internal systems, and consider moving inference to GPU clouds to save cost.
- Balance depends on where your spend is: inference often dominates spend now.
New Chips Free Up Older Capacity
- GPU scarcity eased but remains; new generations free up on-demand capacity for previous chips.
- Increased on-demand supply enables more flexible pricing models like reservations and spot markets on GPU clouds.

