Interconnects

Interviewing Eugene Vinitsky on self-play for self-driving and what else people do with RL

67 snips
Mar 12, 2025
Eugene Vinitsky, a professor at NYU's Civil and Urban Engineering department, dives into the fascinating world of reinforcement learning (RL). He discusses groundbreaking results in self-play for self-driving technology and its implications for future RL applications. The complexity of self-play in multi-agent systems is explored, alongside its surprising link to language model advancements. Eugene shares insights on scaling simulations, the importance of reward design, and the rich potential of AI collaboration, making for a thought-provoking conversation about the future of technology.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ANECDOTE

Language Model "Self-Play" Misconception

  • Language models' "talking to themselves" was misinterpreted as self-play, causing confusion.
  • True self-play, like in AlphaGo, aims for superhuman policies, unlike current language models.
INSIGHT

Self-Play in Driving

  • Self-play allows for superhuman agent development in well-defined games (Go, StarCraft, Dota).
  • Applying self-play to domains like driving, with humans in the loop, is a major challenge.
ANECDOTE

Simulating Human-Level Driving

  • Eugene's team simulated human-level driving without human data, focusing on prediction, planning, and control.
  • Initial simulations showed cars crashing, but they learned to navigate around obstacles and stopped cars.
Get the Snipd Podcast app to discover more snips from this episode
Get the app