Clearer Thinking with Spencer Greenberg

AI Safety and Solutions (with Robert Miles)

29 snips
May 22, 2021
Robert Miles, a science communicator focused on AI safety and alignment, shares insights on the pressing need for AI safety as we advance towards artificial general intelligence (AGI). He discusses the complexity of defining utility functions and the potential existential risks involved. The conversation explores instrumental convergence, the unilateralist's curse, and the challenges of creating AI that aligns with human values. Miles emphasizes the importance of community support in science communication and the necessity for responsible management of AI technology.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Predicting Human Desires

  • Having AI predict human desires through machine learning faces generalization and extrapolation challenges.
  • Out-of-distribution behavior can lead to unexpected outcomes.
ANECDOTE

Adversarial Examples and Goodhart's Law

  • Adversarial examples, like slightly altered panda images misclassified as gibbons, highlight AI's over-optimization.
  • When a measure becomes a target (Goodhart's law), it ceases to be a good measure.
ANECDOTE

Over-Optimization in Human Systems

  • Over-optimization in AI mirrors human systems, like salespeople maximizing revenue at the expense of profit.
  • Or students gaming tests without truly mastering the material.
Get the Snipd Podcast app to discover more snips from this episode
Get the app