
Stephen Casper
Final-year computer science Ph.D. student at MIT in the Algorithmic Alignment Group who authored the report section on technical safeguards, focusing on machine-learning approaches to risk management and monitoring.
Top 3 podcasts with Stephen Casper
Ranked by the Snipd community

73 snips
Feb 10, 2026 • 1h 34min
Inside The Second International AI Safety Report with Writers Stephen Clare and Stephen Casper
Stephen Casper, an MIT Ph.D. student focused on technical safeguards for ML risk management, and Stephen Clare, lead writer of the Second International AI Safety Report and former research manager on frontier AI governance, discuss the report’s scope, jagged capability gains, misuse and systemic risks, the AI development lifecycle, data curation, fine-tuning defenses, post-deployment monitoring and policy priorities for evidence and resilience.

Aug 2, 2024 • 60min
#10: Stephen Casper on Technical and Sociotechnical AI Safety Research
Stephen Casper, a PhD student at MIT specializing in AI safety, dives into the intricacies of AI interpretability and the looming challenges of deceptive alignment. He explains the subtle complexities behind unobservable failures in AI systems, emphasizing the importance of robust evaluations and audits. The discussion also touches on Goodhart's law, illustrating the risks of prioritizing profit over societal well-being, as well as the pressing need for effective governance alongside AI advancements.

Jun 19, 2024 • 2h 42min
Ep 14 - Interp, latent robustness, RLHF limitations w/ Stephen Casper (PhD AI researcher, MIT)
PhD AI researcher Stephen Casper discusses interpretability, robustness, and limitations of AI models. They explore AI safety, risks, research gaps, and advice for early career researchers. Casper emphasizes finding flaws in neural nets, AGI risks, and practical AI applications.


