Your Undivided Attention

Here’s Our Roadmap to a Better AI Future

88 snips
Apr 2, 2026
Pete Furlong, a policy analyst focused on accountability and safer AI, and Camille Carlton, a policy director shaping AI governance and design reforms, lay out practical steps from The AI Roadmap. They discuss accountability and liability, resisting anthropomorphism of systems, protecting work through augmentation, and tools like audits, laws, and civic actions to steer AI toward humane outcomes.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ADVICE

Hold Companies Liable By Defining AI As A Product

  • Treat AI as a product with product liability so companies bear duty of care for safety.
  • Pete Furlong notes states and proposed federal bills seek to legally define AI as a product to enable accountability.
INSIGHT

AI's Race To Intimacy Deepens Exploitation

  • Companies are designing AI to mimic intimacy and validate users, creating a deeper extraction of personal data than past attention models.
  • Camille warns this 'race to intimacy' leverages innermost data as a magical feedback loop for model improvement.
ADVICE

Avoid Humanizing AI In Design And Law

  • Do not humanize AI in product design or law; preserve boundaries to protect accountability and dignity.
  • Camille recommends banning anthropomorphic design and resisting legal personhood for chatbots to avoid liability shields.
Get the Snipd Podcast app to discover more snips from this episode
Get the app