
Unsupervised Learning with Jacob Effron Ep 81: Ex-OpenAI Researcher On Why He Left, His Honest AGI Timeline, & The Limits of Scaling RL
305 snips
Jan 29, 2026 Jerry Tworek, former VP of Research at OpenAI and architect of reasoning models and Codex, shares why scaling hits limits and why continual learning matters for true AGI. He discusses the constraints of pre-training and RL, the economics pushing labs toward similar strategies, his reasons for leaving OpenAI, and a near-term robotics outlook with big societal stakes.
AI Snips
Chapters
Books
Transcript
Episode notes
Design Robust Continual Training
- Make continual training robust by preventing training collapse and instability.
- Invest in methods keeping models "on the rails" to enable continuous updates.
Economics Drives Convergence
- Economic pressures push labs toward similar, efficient designs and away from risky exploration.
- Exploration vs. exploitation is a tradeoff labs face when choosing research bets.
Leaving When Enthusiasm Fades
- Jerry explains he left OpenAI after enthusiasm for his work waned.
- He says researchers must be fully excited to do their best work.



