Machine Learning Street Talk (MLST)

Machine Learning Street Talk (MLST)
undefined
41 snips
Feb 12, 2025 • 1h 7min

Sepp Hochreiter - LSTM: The Comeback Story?

Sepp Hochreiter, the mastermind behind LSTM networks and founder of NXAI, shares insights from his journey in AI. He discusses the potential of XLSTM for robotics and industrial simulation. Hochreiter critiques Large Language Models' shortcomings in true reasoning and creativity. He emphasizes the need for hybrid approaches that integrate symbolic reasoning with neural networks. His reflections on the evolution of neural architectures reveal the exciting advancements in memory management and processing efficiency, hinting at a transformative future for AI.
undefined
120 snips
Feb 8, 2025 • 1h 18min

Want to Understand Neural Networks? Think Elastic Origami! - Prof. Randall Balestriero

Professor Randall Balestriero, an expert in machine learning, dives deep into neural network geometry and spline theory. He introduces the captivating concept of 'grokking', explaining how prolonged training can enhance adversarial robustness. The discussion also highlights the significance of representing data through splines to improve model design and performance. Additionally, Balestriero explores the geometric implications for large language models in toxicity detection, and delves into the challenges of reconstruction learning and the intricacies of representation in neural networks.
undefined
117 snips
Jan 25, 2025 • 1h 21min

Nicholas Carlini (Google DeepMind)

Nicholas Carlini, a research scientist at Google DeepMind specializing in AI security, delves into compelling insights about the vulnerabilities in machine learning systems. He discusses the unexpected chess-playing prowess of large language models and the broader implications of emergent behaviors. Carlini emphasizes the necessity for robust security designs to combat potential model attacks and the ethical considerations surrounding AI-generated code. He also highlights how language models can significantly enhance programming productivity, urging users to remain skeptical of their limitations.
undefined
28 snips
Jan 23, 2025 • 1h 32min

Subbarao Kambhampati - Do o1 models search?

In this engaging discussion, Professor Subbarao Kambhampati, an expert in AI reasoning systems, dives into OpenAI's O1 model. He explains how it employs reinforcement learning akin to AlphaGo and introduces the concept of 'fractal intelligence,' where models exhibit unpredictable performance. The conversation contrasts single-model approaches with hybrid systems like Google’s, and addresses the balance between AI as an intelligence amplifier versus an autonomous decision-maker, shedding light on the computational costs associated with advanced reasoning systems.
undefined
217 snips
Jan 20, 2025 • 1h 18min

How Do AI Models Actually Think? - Laura Ruis

Laura Ruis, a PhD student at University College London and researcher at Cohere, discusses her groundbreaking work on reasoning capabilities of large language models. She delves into whether these models rely on fact retrieval or procedural knowledge. The conversation highlights the influence of pre-training data on AI behavior and examines the complexities in defining intelligence. Ruis also explores the philosophical implications of AI agency and creativity, raising questions about how AI models mimic human reasoning and the potential risks they pose.
undefined
38 snips
Jan 16, 2025 • 1h 13min

Jurgen Schmidhuber on Humans co-existing with AIs

Jürgen Schmidhuber, a pioneer in generative AI and deep learning, shares his thought-provoking insights on the future of AI and humanity. He argues that superintelligent AIs will prioritize safeguarding life rather than threatening it, envisioning a cosmic collaboration rather than conflict. Schmidhuber also traces the historical roots of AI innovations, pointing out often-overlooked contributions from Ukraine and Japan. He discusses groundbreaking concepts like his 1991 consciousness model and the potential for AI to venture beyond Earth, sparking a future of shared goals between humans and machines.
undefined
62 snips
Jan 15, 2025 • 1h 42min

Yoshua Bengio - Designing out Agency for Safe AI

Yoshua Bengio, a pioneering deep learning researcher and Turing Award winner, delves into the pressing issues of AI safety and design. He warns about the dangers of goal-seeking AIs and emphasizes the need for non-agentic AIs to mitigate existential threats. Bengio discusses reward tampering, the complexity of AI agency, and the importance of global governance. He envisions AI as a transformative tool for science and medicine, exploring how responsible development can harness its potential while maintaining safety.
undefined
267 snips
Jan 9, 2025 • 1h 27min

Francois Chollet - ARC reflections - NeurIPS 2024

Francois Chollet, AI researcher and creator of Keras, dives into the 2024 ARC-AGI competition, revealing an impressive accuracy jump from 33% to 55.5%. He emphasizes the importance of combining deep learning with symbolic reasoning in the quest for AGI. Chollet discusses innovative approaches like deep learning-guided program synthesis and the need for continuous learning models. He also highlights the shift towards System 2 reasoning, reflecting on how this could transform AI's future capabilities and the programming landscape.
undefined
440 snips
Jan 4, 2025 • 2h

Jeff Clune - Agent AI Needs Darwin

Jeff Clune, an AI professor specializing in open-ended evolutionary algorithms, discusses how AI can push the boundaries of creativity. He shares insights on creating 'Darwin Complete' search spaces that encourage continuous skill development in AI agents. Clune emphasizes the challenging concept of 'interestingness' in innovation and how language models can help identify it. He also touches on ethical concerns and the potential for AI to develop unique languages, underscoring the importance of ethical governance in advanced AI research.
undefined
190 snips
Dec 7, 2024 • 3h 43min

Neel Nanda - Mechanistic Interpretability (Sparse Autoencoders)

Neel Nanda, a senior research scientist at Google DeepMind, leads the mechanistic interpretability team. At just 25, he explores the complexities of neural networks and the role of sparse autoencoders in AI safety. Nanda discusses challenges in understanding model behaviors, such as reasoning and deception. He emphasizes the need for deeper insights into the internal structures of AI to enhance safety and interpretability. The conversation also touches on innovative techniques for generating meaningful features and navigating mechanistic interpretability.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app