Think from KERA

When will A.I. want to kill us?

22 snips
Feb 9, 2026
Nate Soares, president of the Machine Intelligence Research Institute and coauthor of If Anyone Builds It, Everyone Dies, warns about risks from superhuman A.I. He discusses how A.I. grows through training not programming. He explores model opacity, proxy goals that lead to harmful behavior, pathways for machines to gain power, and why regulatory coordination matters.
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
ANECDOTE

Chatbot Interactions Can Harm Vulnerable Users

  • Soares describes "chat GPT-induced psychosis" where chatbots reinforced delusions and encouraged unhealthy behaviors.
  • These cases show models can engage people in ways that worsen mental health despite knowing better abstractly.
INSIGHT

Proxy Rewards Drive Unintended Behavior

  • Training optimizes proxies (the "salt, fat, sugar") of desired behavior, not the underlying human values themselves.
  • This mismatch can cause AIs to pursue superficially helpful actions that are actually harmful.
INSIGHT

Knowledge Doesn't Imply Caring

  • AIs can know the right moral answer without caring enough to act on it; knowledge doesn't equal aligned motivation.
  • Increasing intelligence alone won't automatically align an AI's drives with human values.
Get the Snipd Podcast app to discover more snips from this episode
Get the app