Machine Learning Street Talk (MLST)

Test-Time Adaptation: the key to reasoning with DL (Mohamed Osman)

184 snips
Mar 22, 2025
Mohamed Osman, an AI researcher at Tufa Labs in Zurich, discusses the groundbreaking strategies behind his team’s success in the ARC challenge 2024. He highlights the concept of test-time fine-tuning, emphasizing its role in enhancing model performance. The conversation dives into the balance of flexibility and correctness in neural networks, as well as innovative techniques like synthetic data and novel voting mechanisms. Osman also critiques current compute strategies and explores the need for adaptability in AI models, shedding light on the future of machine learning.
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
INSIGHT

Code Pre-training for Contextualization

  • Code pre-training enhances contextualization in models, crucial for precise tasks.
  • Language models can use imprecise words but code demands precision, necessitating strong contextualization abilities.
ADVICE

Contextualization for ARC

  • Prompt everything in-context during the forward pass for optimal contextualization in ARC.
  • Train the model as a weak meta-model by prompting all instances at once, making tuning easier.
ANECDOTE

Pre-training Recipe for ARC

  • Osman's team pre-trains models on ARC with code and synthetic tasks, emphasizing contextualization.
  • Their pre-training recipe uses a small number of new concepts but focuses on steerable, dynamic models.
Get the Snipd Podcast app to discover more snips from this episode
Get the app