
Machine Learning Street Talk (MLST) #55 Self-Supervised Vision Models (Dr. Ishan Misra - FAIR).
21 snips
Jun 21, 2021 Dr. Ishan Misra, a prolific Research Scientist at Facebook AI Research, dives into the world of self-supervised vision models. He discusses groundbreaking papers like DINO and Barlow Twins, addressing how these innovative approaches reduce the need for human supervision in visual learning. Ishan explores the nuances of neural networks, object recognition challenges, and the philosophical implications of AI's common sense knowledge. Plus, he compares self-supervised models with semi-supervised techniques, showcasing the advancements in harnessing human knowledge for enhanced machine learning.
AI Snips
Chapters
Books
Transcript
Episode notes
Semantics of Similarity
- Self-supervised models learn semantics of similarity, not language-induced semantics.
- Attaching language to these similarity representations is needed for communication.
Dataset Bias
- Datasets like ImageNet are biased towards object-centric images.
- Random images would likely break current self-supervised systems.
Richer Labeling
- Explore relationships between objects by asking how images are related or unrelated.
- This richer labeling encodes human knowledge and object properties.


























