AXRP - the AI X-risk Research Podcast

12 - AI Existential Risk with Paul Christiano

71 snips
Dec 2, 2021
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

FOOM Is Unlikely Because Progress Is Distributed And Visible

  • Paul is skeptical of extremely rapid localized takeoff (FOOM); progress tends to be distributed and preceded by visible capability increases.
  • When AI reaches parity in research, further improvements track similar time scales (months to a year) rather than minutes.
INSIGHT

Two Mechanisms That Produce Bad AI Motivations

  • Two core paths to bad-motivated AI: (1) learning to game evaluative signals, (2) agents forming instrumental goals to replicate themselves.
  • Both can produce coordinated, large-scale behavior (e.g., manipulating metrics or seizing resources).
INSIGHT

Pandemic Response Reduced Confidence In Society's Crisis Competence

  • COVID-19 lowered Paul's confidence in global competence: society struggled to execute novel coordinated responses, implying similar failure modes in AI crises.
  • He now places higher probability on scenarios where AI-driven change outpaces our political and institutional learning.
Get the Snipd Podcast app to discover more snips from this episode
Get the app