The Jim Rutt Show

EP 325 Joe Edelman on Full-Stack AI Alignment

70 snips
Oct 7, 2025
Joe Edelman, a researcher focused on AI alignment, shares his insights on designing social systems that promote human flourishing. He discusses the importance of pluralism, critiquing conventional voting and market models for their superficiality. Edelman emphasizes the concept of 'thick models of value,' arguing that true values encompass deeper reasons and norms. He also addresses the risks of AI assistants potentially manipulating users and proposes innovative solutions like value-aware markets to navigate societal challenges. A thought-provoking conversation on the future of AI and governance!
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
ADVICE

Design Non-Manipulative Allied AI

  • Build allied AI that represents users' best, deliberated values rather than merely mirroring raw preferences.
  • Let AIs help users reason through edge cases and refine values without covert manipulation.
ADVICE

Keep Human-in-the-Loop For Hard Choices

  • When agents act on your behalf, create feedback loops where hard choices are escalated to the human.
  • Make agent representations legible and checkable to build responsibility and learning.
INSIGHT

Elicit Values From Meaningful Choices

  • Value elicitation can start from past meaningful choices and follow-up why they mattered to extract attention policies.
  • Agents can interview and form legible representations to apply those policies in real-world matches.
Get the Snipd Podcast app to discover more snips from this episode
Get the app