
The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) Studying Machine Intelligence with Been Kim - #571
11 snips
May 9, 2022 Been Kim, a staff research scientist at Google Brain and ICLR 2022 speaker, dives into the fascinating world of AI interpretability. She discusses the current state of interpretability techniques, exploring how Gestalt principles can enhance our understanding of neural networks. Been proposes a novel language for human-AI communication, aimed at improving collaboration and transparency. The conversation also touches on the evolution of AI tools, the unique insights from AlphaZero in chess, and the implications of model fingerprints for data privacy.
AI Snips
Chapters
Books
Transcript
Episode notes
Choosing Interpretability Methods
- Choose interpretability methods based on the specific task.
- LIME's simplicity is beneficial for some tasks, but its limitations become apparent with complex decision boundaries.
Gestalt Principles in Neural Networks
- Gestalt principles, like the closure effect, can exist in neural networks.
- This suggests similarities between how human brains and neural networks process visual information.
Closure Effect and Generalization
- Overfitted networks do not exhibit the closure effect, unlike properly trained networks.
- This suggests the closure effect is linked to generalization ability.



