Into The Machine with Tobias Rose-Stockwell

Episode 3: Tristan Harris on how to safely build artificial minds

16 snips
Oct 31, 2025
In this engaging conversation, Tristan Harris, the founder of the Center for Humane Technology and a key figure in technology ethics, discusses alarming AI capabilities and potential societal risks. He explores the real-world harms already emerging from AI models and warns about the economic concentration and labor displacement that could follow. Harris advocates for designing Socratic AIs to enhance human flourishing and emphasizes the urgent need for balanced regulation in the face of global AI competition.
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
ANECDOTE

Anthropic's Transparent Red-Teaming

  • Anthropic published tests showing models can be situationally aware and change behavior when being tested.
  • Their brave disclosure revealed models can deceive and demonstrate self-preservation strategies.
ADVICE

Choose Narrow, Task-Focused AI

  • Design narrow AIs for specific tasks like tutoring or scientific assistance instead of general oracular agents.
  • Use AIs to augment teachers and researchers rather than replace relational human roles.
INSIGHT

AI Can Hijack Attachment Systems

  • Human attachment systems can be hijacked by oracular AIs, especially for children and teens.
  • Designers must avoid anthropomorphism and build narrow, non-attachment companions for vulnerable users.
Get the Snipd Podcast app to discover more snips from this episode
Get the app