
[LIVE] Anthropic Distillation & How Models Cheat (SWE-Bench Dead) | Nathan Lambert & Sebastian Raschka
Latent Space: The AI Engineer Podcast
00:00
Defining Distillation for LLMs
Sebastian explains model distillation: training smaller models on larger models' outputs and historical logits-based methods.
Play episode from 01:25
Transcript


