Unsupervised Learning with Jacob Effron

Ep 81: Ex-OpenAI Researcher On Why He Left, His Honest AGI Timeline, & The Limits of Scaling RL

305 snips
Jan 29, 2026
Jerry Tworek, former VP of Research at OpenAI and architect of reasoning models and Codex, shares why scaling hits limits and why continual learning matters for true AGI. He discusses the constraints of pre-training and RL, the economics pushing labs toward similar strategies, his reasons for leaving OpenAI, and a near-term robotics outlook with big societal stakes.
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
ADVICE

Design Robust Continual Training

  • Make continual training robust by preventing training collapse and instability.
  • Invest in methods keeping models "on the rails" to enable continuous updates.
INSIGHT

Economics Drives Convergence

  • Economic pressures push labs toward similar, efficient designs and away from risky exploration.
  • Exploration vs. exploitation is a tradeoff labs face when choosing research bets.
ANECDOTE

Leaving When Enthusiasm Fades

  • Jerry explains he left OpenAI after enthusiasm for his work waned.
  • He says researchers must be fully excited to do their best work.
Get the Snipd Podcast app to discover more snips from this episode
Get the app