The Data Exchange with Ben Lorica

World Models Are Here—But It’s Still the GPT-2 Phase

Mar 19, 2026
Jeff Hawke, CTO of Odyssey, builds general-purpose world models that generate interactive visual simulations from images or text. He explains how continuous video-like models are trained, early use cases like games and robotics, compute and latency challenges, stability limits on long runs, and the path toward scalable, real-time and on-device deployments.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

World Models Are In The GPT-2 Phase

  • World models are early and exploratory—Jeff Hawke compares the current period to the GPT-2 era of LLMs.
  • He expects many experimental use cases (games, retail, live events, robotics) before mass commercialization.
ADVICE

Use Transformers But Train For Dynamics

  • Use transformer architectures as the backbone but design training regimes for dynamics rather than static generation.
  • Jeff Hawke stresses transformers underpin frontier models, but framing and training determine whether you learn world dynamics.
INSIGHT

Current World Models Struggle With Long Rollouts

  • Major limitations today are short-range prediction and instability over long rollouts; state-of-the-art contiguous prediction moved from 15–30 seconds (mid‑2025) to roughly 1–2 minutes.
  • Instability appears as visual degradation or 'psychedelic' noise when models go out of distribution.
Get the Snipd Podcast app to discover more snips from this episode
Get the app