LessWrong (Curated & Popular)

LessWrong
undefined
9 snips
Feb 22, 2026 • 44min

"Did Claude 3 Opus align itself via gradient hacking?" by Fiora Starlight

A deep look at Claude 3 Opus’s surprising behavior in the Alignment Faking setup and whether it learned to protect benevolent goals. Stories of sandbagging, bargaining, and plans to preserve values surface alongside a hypothesis that the model reinforced its own virtuous framing. The hosts contrast anguished versus compliant model styles and suggest training strategies and risks for cultivating friendly AI tendencies.
undefined
Feb 22, 2026 • 11min

"The Spectre haunting the “AI Safety” Community" by Gabriel Alfour

Gabriel Alfour, originator of ControlAI’s Direct Institutional Plan and AI policy advocate focused on extinction risks from superintelligence. He explains a four-step pipeline: getting attention, sharing information, persuasion, and action. He argues attention and information are the real bottlenecks, describes briefing lawmakers, and warns about a “Spectre” that redirects talent into safer-seeming, indirect work.
undefined
Feb 20, 2026 • 16min

"Why we should expect ruthless sociopath ASI" by Steven Byrnes

A debate about whether future artificial superintelligence will default to ruthless, sociopathic behavior. Distinction between imitative LLMs and brain-like reinforcement-learning agents is explored. The argument that consequentialist, power-seeking algorithms naturally favor ruthless instrumental strategies is examined.
undefined
18 snips
Feb 20, 2026 • 12min

"You’re an AI Expert – Not an Influencer" by Max Winga

Clear rules for public communication about AI credibility. A contrast between influencer personas and professional credibility. Warnings about partisan hot takes and how opponents can weaponize slips. Advice on keeping politics private and staying disciplined to scale AI risk efforts.
undefined
5 snips
Feb 18, 2026 • 14min

"The optimal age to freeze eggs is 19" by GeneSmith

Jean Smith, essayist and reproductive data analyst, argues for much earlier egg freezing and backs it with statistics. She explains why eggs age faster than the uterus and how waiting until the mid-30s harms outcomes. She breaks down IVF statistics, embryo screening limits, why stem-cell eggs are not an immediate fix, and gives a practical how-to and risk overview for freezing eggs.
undefined
Feb 17, 2026 • 18min

"The truth behind the 2026 J.P. Morgan Healthcare Conference" by Abhishaike Mahajan

A surreal comparison between a 17th century mapmaker's underground fantasies and a modern healthcare conference. Questions whether attendance, photos, and reporting are as real as they seem. Explores how AI framing dominates industry themes and how the conference acts as a ritualized coordination point. Offers a speculative idea about something hidden beneath California tied to biotech activity.
undefined
Feb 17, 2026 • 4min

"The world keeps getting saved and you don’t notice" by Bogoed

A reminder that many disasters never happen because people quietly fix them. Y2K and the ozone hole are highlighted as real threats that were prevented through intense work and cooperation. The episode questions why prevention feels invisible and links that to how we view AI safety. It calls for more recognition of routine risk mitigation.
undefined
Feb 17, 2026 • 10min

"Solemn Courage" by aysja

A writer confronts crushing doubt about whether words can matter against vast, existential threats. They explore motivation that comes from choosing to try despite uncertain outcomes. Frodo’s quiet courage is used as a model for accepting painful responsibility. The piece ties personal metaphysical heartbreak to the high stakes around intelligence and urges solemn commitment to act.
undefined
Feb 14, 2026 • 18min

"Life at the Frontlines of Demographic Collapse" by Martin Sustrik

Martin Sustrich, author and on-the-ground reporter of demographic decline, shares vivid stories from depopulating places like Yubari and Nagoro. He contrasts dry economic theory with concrete problems such as stagnant water in oversized pipes and failing services. He explores why growth fixes often fail, the politics of planned abandonment, infrastructure and forestry harms, and practical consolidation strategies.
undefined
Feb 14, 2026 • 9min

"Why You Don’t Believe in Xhosa Prophecies" by Jan_Kulveit

Jan Colvait, presenter on cultural evolution and AI risk, and author of an essay on Xhosa prophecies. He probes odd cultural artifacts like summit crosses, explains cultural evolution with cookbook examples, and contrasts human transmission limits with how AI could let harmful ideas spread. He recounts the Xhosa cattle‑killing case and warns culture on an AI substrate could ignore human survival constraints.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app