The MAD Podcast with Matt Turck

What’s Next for AI? OpenAI’s Łukasz Kaiser (Transformer Co-Author)

204 snips
Nov 26, 2025
Łukasz Kaiser, a leading researcher at OpenAI and co-author of the influential 'Attention Is All You Need' paper, delves into the latest advancements in AI, including GPT-5.1. He explains the steady exponential growth in AI capabilities, the significance of reasoning models, and how modern chat models utilize tools to enhance their performance. Kaiser also discusses the messy reality of engineering challenges, the future of pre-training, and why even cutting-edge models can struggle with simple logic puzzles. His journey from academia to shaping AI innovation offers a personal touch.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Attention And Practical Engineering Matter

  • 'Attention' lets the model look back at relevant past tokens; Transformer novelty was self-attention plus architectural and optimization tweaks.
  • Łukasz highlights engineering tricks (warmup, optimizers) that made the architecture trainable at scale.
INSIGHT

Pre-Training Didn't Die—Economics Shifted

  • Pre-training still scales smoothly and reduces loss with more compute, but economics shifted focus to smaller, cheaper models for large user bases.
  • Distillation makes huge pre-trained models valuable as teachers for cheaper production models.
INSIGHT

We Can Open Some Black Boxes—Not All

  • Interpretability has advanced: sparse models and circuit analyses reveal mechanisms, but full understanding of largest models remains limited.
  • Łukasz notes progress on circuits and sparsity but accepts practical limits of comprehending very complex systems.
Get the Snipd Podcast app to discover more snips from this episode
Get the app