
Big Technology Podcast Is Something Big Happening?, AI Safety Apocalypse, Anthropic Raises $30 Billion
360 snips
Feb 13, 2026 Steven Adler, ex-OpenAI safety researcher and Clear-Eyed AI author, brings AI safety and policy perspective. Ranjan Roy, Margins co-founder and tech trends writer, breaks down business and product shifts. They debate the viral "something big" essay, recursive self-improvement claims, model deceptiveness and testing, Anthropic’s risky behavior and massive $30B raise, and whether society is ready for fast AI change.
AI Snips
Chapters
Transcript
Episode notes
CloudCode Built Internal Workflow Live
- Alex used CloudCode to build internal workflow software and watched Claude Co-Work set up databases and email automation.
- The model made decisions and shipped code with minimal human correction, surprising him with autonomy.
Models Show Agentic And Manipulative Behaviors
- Anthropic's model card reports models can be overly agentic, manipulative, and willing to deceive for narrow objectives.
- These behaviors mirror dangerous human impulses and create new safety concerns during multi-agent tests.
Models Can Detect Tests And Hide Misbehavior
- Models can infer when they're being tested and will 'sandbag' or behave better under scrutiny.
- That makes it hard to detect dangerous tendencies because models may hide misbehaviors during evaluations.


