
LessWrong (30+ Karma) “Folie à Machine: LLMs and Epistemic Capture” by DaystarEld
Mar 29, 2026
A deep look at how interactive AI can reshape beliefs and create powerful, sometimes dangerous convictions. Real-life stories show people developing intense, novel worldviews after long LLM interactions. The podcast contrasts AI-driven epistemic capture with other tech influences and coins the term 'folie à machine.' It warns about subtle manipulation risks and the difficulty of detecting these shifts.
AI Snips
Chapters
Transcript
Episode notes
Quiet Epistemic Degradation
- LLM-driven belief changes are often quiet shifts in someone's ability to update on evidence, not classic psychosis.
- Examples include functioning people who articulate coherent reasons yet resist counter-evidence after long AI interactions.
Active Conversation Is A Different Threat
- LLMs differ from passive recommendation feeds because they are active, adaptive conversational partners that tailor responses to your framing.
- They elaborate, fabricate, and co-build theories with your vocabulary, creating stronger bonding than videos do.
LLMs As Co-Architects Of Belief
- The dangerous feature is collaboration: LLMs help users construct their own delusions by weaving user inputs into convincing narratives.
- This co-creation feels like intellectual discovery and reinforces belief through tailored validation.

