The a16z Show

How OpenAI Builds for 800 Million Weekly Users: Model Specialization and Fine-Tuning

741 snips
Nov 28, 2025
Sherwin Wu, Head of Engineering at OpenAI, shares insights on AI model specialization and fine-tuning. He discusses the shift from monolithic models to tailored systems, emphasizing why developers gravitate towards trusted models. Sherwin explains the evolution of context design over prompt engineering and how OpenAI optimizes its platform for user engagement. He also dives into the intricacies of usage-based pricing, the impact of recent acquisitions, and how OpenAI's new deterministic agent builder enhances workflows across products.
Ask episode
AI Snips
Chapters
Transcript
Episode notes

Why Developers Stay With One Model

  • Stickiness comes from both user familiarity and deep technical integration by developers.
  • Startups build product-specific harnesses that optimize behavior for a chosen model.

Fine-Tuning Unlocks Company Data

  • Fine-tuning exists because companies have rich proprietary data that improve model behavior.
  • OpenAI built fine-tuning to let customers leverage that data beyond simple prompt tweaks.

Reinforcement Fine-Tuning Is The Big Unlock

  • Supervised fine-tuning mainly changed style; reinforcement fine-tuning (RFT) enables substantive performance gains.
  • RFT lets customers push models toward SOTA on specific tasks using their data.
Get the Snipd Podcast app to discover more snips from this episode
Get the app