

Machine Learning Street Talk (MLST)
Machine Learning Street Talk (MLST)
Welcome! We engage in fascinating discussions with pre-eminent figures in the AI field. Our flagship show covers current affairs in AI, cognitive science, neuroscience and philosophy of mind with in-depth analysis. Our approach is unrivalled in terms of scope and rigour – we believe in intellectual diversity in AI, and we touch on all of the main ideas in the field with the hype surgically removed. MLST is run by Tim Scarfe, Ph.D (https://www.linkedin.com/in/ecsquizor/) and features regular appearances from MIT Doctor of Philosophy Keith Duggar (https://www.linkedin.com/in/dr-keith-duggar/).
Episodes
Mentioned books

41 snips
Feb 12, 2025 • 1h 7min
Sepp Hochreiter - LSTM: The Comeback Story?
Sepp Hochreiter, the mastermind behind LSTM networks and founder of NXAI, shares insights from his journey in AI. He discusses the potential of XLSTM for robotics and industrial simulation. Hochreiter critiques Large Language Models' shortcomings in true reasoning and creativity. He emphasizes the need for hybrid approaches that integrate symbolic reasoning with neural networks. His reflections on the evolution of neural architectures reveal the exciting advancements in memory management and processing efficiency, hinting at a transformative future for AI.

120 snips
Feb 8, 2025 • 1h 18min
Want to Understand Neural Networks? Think Elastic Origami! - Prof. Randall Balestriero
Professor Randall Balestriero, an expert in machine learning, dives deep into neural network geometry and spline theory. He introduces the captivating concept of 'grokking', explaining how prolonged training can enhance adversarial robustness. The discussion also highlights the significance of representing data through splines to improve model design and performance. Additionally, Balestriero explores the geometric implications for large language models in toxicity detection, and delves into the challenges of reconstruction learning and the intricacies of representation in neural networks.

117 snips
Jan 25, 2025 • 1h 21min
Nicholas Carlini (Google DeepMind)
Nicholas Carlini, a research scientist at Google DeepMind specializing in AI security, delves into compelling insights about the vulnerabilities in machine learning systems. He discusses the unexpected chess-playing prowess of large language models and the broader implications of emergent behaviors. Carlini emphasizes the necessity for robust security designs to combat potential model attacks and the ethical considerations surrounding AI-generated code. He also highlights how language models can significantly enhance programming productivity, urging users to remain skeptical of their limitations.

28 snips
Jan 23, 2025 • 1h 32min
Subbarao Kambhampati - Do o1 models search?
In this engaging discussion, Professor Subbarao Kambhampati, an expert in AI reasoning systems, dives into OpenAI's O1 model. He explains how it employs reinforcement learning akin to AlphaGo and introduces the concept of 'fractal intelligence,' where models exhibit unpredictable performance. The conversation contrasts single-model approaches with hybrid systems like Google’s, and addresses the balance between AI as an intelligence amplifier versus an autonomous decision-maker, shedding light on the computational costs associated with advanced reasoning systems.

217 snips
Jan 20, 2025 • 1h 18min
How Do AI Models Actually Think? - Laura Ruis
Laura Ruis, a PhD student at University College London and researcher at Cohere, discusses her groundbreaking work on reasoning capabilities of large language models. She delves into whether these models rely on fact retrieval or procedural knowledge. The conversation highlights the influence of pre-training data on AI behavior and examines the complexities in defining intelligence. Ruis also explores the philosophical implications of AI agency and creativity, raising questions about how AI models mimic human reasoning and the potential risks they pose.

38 snips
Jan 16, 2025 • 1h 13min
Jurgen Schmidhuber on Humans co-existing with AIs
Jürgen Schmidhuber, a pioneer in generative AI and deep learning, shares his thought-provoking insights on the future of AI and humanity. He argues that superintelligent AIs will prioritize safeguarding life rather than threatening it, envisioning a cosmic collaboration rather than conflict. Schmidhuber also traces the historical roots of AI innovations, pointing out often-overlooked contributions from Ukraine and Japan. He discusses groundbreaking concepts like his 1991 consciousness model and the potential for AI to venture beyond Earth, sparking a future of shared goals between humans and machines.

62 snips
Jan 15, 2025 • 1h 42min
Yoshua Bengio - Designing out Agency for Safe AI
Yoshua Bengio, a pioneering deep learning researcher and Turing Award winner, delves into the pressing issues of AI safety and design. He warns about the dangers of goal-seeking AIs and emphasizes the need for non-agentic AIs to mitigate existential threats. Bengio discusses reward tampering, the complexity of AI agency, and the importance of global governance. He envisions AI as a transformative tool for science and medicine, exploring how responsible development can harness its potential while maintaining safety.

267 snips
Jan 9, 2025 • 1h 27min
Francois Chollet - ARC reflections - NeurIPS 2024
Francois Chollet, AI researcher and creator of Keras, dives into the 2024 ARC-AGI competition, revealing an impressive accuracy jump from 33% to 55.5%. He emphasizes the importance of combining deep learning with symbolic reasoning in the quest for AGI. Chollet discusses innovative approaches like deep learning-guided program synthesis and the need for continuous learning models. He also highlights the shift towards System 2 reasoning, reflecting on how this could transform AI's future capabilities and the programming landscape.

440 snips
Jan 4, 2025 • 2h
Jeff Clune - Agent AI Needs Darwin
Jeff Clune, an AI professor specializing in open-ended evolutionary algorithms, discusses how AI can push the boundaries of creativity. He shares insights on creating 'Darwin Complete' search spaces that encourage continuous skill development in AI agents. Clune emphasizes the challenging concept of 'interestingness' in innovation and how language models can help identify it. He also touches on ethical concerns and the potential for AI to develop unique languages, underscoring the importance of ethical governance in advanced AI research.

190 snips
Dec 7, 2024 • 3h 43min
Neel Nanda - Mechanistic Interpretability (Sparse Autoencoders)
Neel Nanda, a senior research scientist at Google DeepMind, leads the mechanistic interpretability team. At just 25, he explores the complexities of neural networks and the role of sparse autoencoders in AI safety. Nanda discusses challenges in understanding model behaviors, such as reasoning and deception. He emphasizes the need for deeper insights into the internal structures of AI to enhance safety and interpretability. The conversation also touches on innovative techniques for generating meaningful features and navigating mechanistic interpretability.


