80,000 Hours Podcast

#235 – Ajeya Cotra on whether it’s crazy that every AI company’s safety plan is ‘use AI to make AI safe’

250 snips
Feb 17, 2026
Ajeya Cotra, AI researcher and strategist who forecasts timelines and models AI risk, explores whether using AI to make AI safe is sensible. She contrasts gradual versus explosive takeoffs. She outlines feedback loops that could speed progress. She discusses what a crunch-time redirection of AI labor would require and why companies might not cooperate.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ADVICE

Redirect Compute To Defenses During Crunch

  • If an intelligence explosion begins, redirect AI labour from capability racing to defenses: alignment, biodefense, cyberdefense, and coordination.
  • Build plans to rapidly reallocate compute, talent, and funding within a short window (months to a year).
INSIGHT

'Use AI To Make AI Safe' Is Plausible But Precarious

  • Using AI to solve AI problems is historically plausible but uniquely risky because the window to redirect effort may be very short.
  • Success hinges on noticing the window and tilting AI labor toward protective work quickly.
ADVICE

Build Scaffolds Before Handing AIs Power

  • Prioritize control, interpretability, and alignment scaffolding so we can trust AI outputs enough to rely on them for safety tasks.
  • Treat building setups that bottleneck runaway capability as a key research objective now.
Get the Snipd Podcast app to discover more snips from this episode
Get the app