80,000 Hours Podcast

#156 – Markus Anderljung on how to regulate cutting-edge AI models

40 snips
Jul 10, 2023
Markus Anderljung, Head of Policy at the Centre for the Governance of AI, dives into the complex world of AI governance. He discusses the urgent need for regulations on advanced AI, including self-replicating models and the risk of dangerous capabilities. Topics range from the challenges of deploying AI safely to the potential for regulatory capture by the industry. Anderljung emphasizes the importance of proactive measures and international cooperation to ensure accountability and safety in AI development, making this conversation pivotal for anyone interested in the future of technology.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ANECDOTE

LLM Emergent Capabilities

  • Large language models (LLMs) struggle with arithmetic, improving suddenly after figuring out the 'math'.
  • This exemplifies emergent capabilities, making risk prediction difficult.
ANECDOTE

GPT-3's Coding Surprise

  • GPT-3's unexpected coding ability, learned from internet text, led to specialized code models.
  • This highlights how emergent capabilities can redirect AI development.
INSIGHT

Proliferation Problem

  • AI models' dangerous capabilities can proliferate rapidly through replication and theft.
  • Regulation must address this by considering earlier intervention.
Get the Snipd Podcast app to discover more snips from this episode
Get the app