EP39: White House AI Executive Order, The Bletchley Declaration & Adversarial AI Attacks
Nov 2, 2023
The podcast discusses the White House's executive order on regulating AI, the Bletchley Declaration, adversarial AI attacks, future of AI computing, and leaked prompts powering ChatGPT's multi-tool mode. They also introduce hilarious new merch and discuss key insights on the future of language models.
01:06:59
forum Ask episode
web_stories AI Snips
view_agenda Chapters
auto_awesome Transcript
info_circle Episode notes
insights INSIGHT
Risk Framing Often Overreaches
The executive order sometimes frames risks in exaggerated, low-probability terms (e.g., easy biological threats).
Some risks described are already possible without advanced AI and may be framed as AI-enabled for emphasis.
volunteer_activism ADVICE
Adopt Risk-Based AI Use In Government
Embrace AI in government with risk-aware frameworks instead of blanket bans.
Infuse AI talent across agencies and provide safe data access to foster innovation.
volunteer_activism ADVICE
Use AI To Harden Cybersecurity
Include AI-specific checks into cybersecurity frameworks and penetration testing.
Use AI to proactively find and patch vulnerabilities rather than only enabling attackers.
Get the Snipd Podcast app to discover more snips from this episode
Join our Discord: https://discord.gg/TRrgAyeM Buy the merch: https://www.thisdayinaimerch.com/
This week the AI guys unpack the White House's sweeping executive order on regulating AI - will this lead to the death of open-source models? They also discuss the vague and fluffy Bletchley Declaration signed by world leaders, why Geoffrey Hinton just won't stop fearmongering, and introduce some hilarious new merch including a life-size shower curtain! Tune in for hot takes on the AI ethics debate, prompt engineering tricks, and key insights on the future of language models.
CHAPTERS: ===== 00:00 - King Charles on AI (Cold Open) 00:20 - Thoughts on White House AI Executive Order 23:09 - The Bletchley Declaration & AI Safety Summit 38:04 - LoRA Fine-tuning Efficiently Undoes Safety Training in Llama 2 & They Killed Tay! 48:34 - Adversarial Attacks and Defenses in Large Language Models: Old and New Threats Paper 51:51 - Mike proposes What The Future of AI Computing Might Look Like 55:00 - Leaked: The Secret Prompt Powering ChatGPT's New Multi-Tool Mode (and How to Hack It)
1:01:39 - Anthropic Have Raised More Billions & Our Merch Store!