LessWrong (30+ Karma)

“There should be $100M grants to automate AI safety” by Marius Hobbhahn

Apr 3, 2026
Marius Hobbhahn, author and Apollo Research affiliate, proposes massive grants to scale automated AI-safety work. He urges urgent, large-scale funding and a grant model that ramps to $100M+ budgets for automated safety pipelines. He outlines concrete areas like monitoring, automated red-teaming, white-box auditing, propensity evaluations, and automated conceptual alignment research.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ADVICE

Make Large Grants Conditional On Public Benefit

  • Require public benefit and publication when a grantee proves they can spend $100M meaningfully on safety.
  • Options include open-sourcing the pipeline, collaborating with labs, joining an AGI lab, or turning into an AGI-safety for-profit under publication conditions.
INSIGHT

Explicit Programs Attract Entrepreneurial Talent

  • Announce explicit grant programs because conservatism among funders means entrepreneurs won't attempt ambitious scaling without clear commitments.
  • Hobbhahn expects entrepreneurial talent needs visible incentives to choose safety-first scaling projects.
INSIGHT

Goodharting Can Be A Useful Failure Signal

  • Goodharting is a real risk but observing metric failure is informative evidence to stop that program.
  • Hobbhahn suggests robust metrics help, and failed metrics signal the need to halt particular scaling attempts.
Get the Snipd Podcast app to discover more snips from this episode
Get the app