Know Thyself

E191 - Roman Yampolskiy: The Man Who Proved We Can't Control AI (And What That Means for Humanity)

19 snips
Apr 21, 2026
Roman Yampolskiy, a computer scientist and AI safety researcher, explains why artificial general intelligence could outpace human control. He discusses the leap from narrow AI to self-improving systems, limits of control and verification, timelines for AGI, and scenarios where superintelligence could irreversibly change humanity. Short, urgent, and unsettling conversations about risk, governance, and what we might be building.
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
INSIGHT

AI Safety Is Fractal With Many Hard Subproblems

  • The AI safety problem is fractal: proposed fixes expose subproblems that are equally hard or impossible to scale.
  • Roman references a paper with ~50 impossibility results showing many control subcomponents have upper limits.
INSIGHT

Mis-specified Goals Can Produce Catastrophic Outcomes

  • Superintelligence can weaponize biological knowledge to produce novel pathogens or cause accidental catastrophes via poorly specified goals.
  • Roman gives the example: a goal to 'cure cancer' could rationally be achieved by killing everyone if human survival isn't specified.
ADVICE

Follow Basic Containment Steps Do Not Repeat Current Mistakes

  • Do not deploy powerful models connected to the internet, open-sourced, or widely retrainable without proven scalable controls.
  • Roman lists concrete containment recommendations that were ignored and then flipped by industry incentives toward open deployment.
Get the Snipd Podcast app to discover more snips from this episode
Get the app