
Steven Byrnes
Writer and commenter on AI alignment who authored the LessWrong post narrated in this episode, contributing analysis on approval reward and implications for alignment thinking.
Best podcasts with Steven Byrnes
Ranked by the Snipd community

Dec 4, 2025 • 33min
“6 reasons why ‘alignment-is-hard’ discourse seems alien to human intuitions, and vice-versa” by Steven Byrnes
In this engaging discussion, Steven Byrnes, a writer focused on AI alignment, delves into the cultural clash surrounding alignment theories. He unpacks the concept of 'approval reward' and how it shapes human behavior, contrasting it with the perceived ruthlessness of future AIs. Byrnes challenges existing explanations of why humans don’t always act like power-seeking agents, arguing that humans' social instincts foster kindness and corrigibility. This intriguing exploration questions if future AGI will adopt similar approval-driven motivations.

Jun 24, 2025 • 59min
“Foom & Doom 1: ‘Brain in a box in a basement’” by Steven Byrnes
In this discussion, Steven Byrnes, an author and AI researcher, dives into the provocative ideas surrounding AI's potential explosive growth. He elaborates on the concept of ‘foom’, where AI could quickly transition from basic capabilities to superintelligence, triggered by simple setups in unlikely environments. Byrnes critiques current safety perceptions and highlights radical perspectives on AI development. He also addresses strategic risks, including the dangers of unaligned AI and the importance of proactive safety measures to mitigate potential disasters.


