The MAD Podcast with Matt Turck

The End of GPU Scaling? Compute & The Agent Era — Tim Dettmers (Ai2) & Dan Fu (Together AI)

122 snips
Jan 22, 2026
Tim Dettmers, an assistant professor at Carnegie Mellon University, and Dan Fu, an assistant professor at UC San Diego, dive deep into the future of AGI. They debate the limitations of current hardware versus the untapped potential of efficient utilization. Tim warns of physical constraints like the von Neumann bottleneck, while Dan emphasizes better performance through optimized kernels. The conversation also reveals how agents can enhance productivity, with practical advice on leveraging them effectively for work automation and innovation in AI architectures.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Models Lag Behind Hardware Reality

  • Many current models were trained on older clusters and underutilized hardware.
  • Dan estimates up to ~100x more effective compute is already available when accounting for new chips and utilization gains.
INSIGHT

Pretraining vs Post-Training Tradeoff

  • Pre-training builds general capability while post-training tailors useful, product-ready skills.
  • Post-training and product feedback matter more for real-world usefulness than raw scale alone.
ANECDOTE

When Agents Beat The 'Final Boss'

  • Dan describes the moment agents crossed a practical threshold for him in June 2025.
  • He used agents to write complex GPU kernels and became multiple-times more productive overnight.
Get the Snipd Podcast app to discover more snips from this episode
Get the app