Software Engineering Radio - the podcast for professional software developers

SE Radio 715: Sahaj Garg on Designing for Ambiguity in Human Input

5 snips
Apr 8, 2026
Sahaj Garg, co-founder and CTO of Wispr and former AI engineer with ML research, discusses designing for ambiguity in human input. He explores categories like lexical, syntactic, and voice-specific ambiguity. Practical topics include using context and personalization, dataset construction and annotation, inference tradeoffs, revealed preferences from user edits, and subtle ways to communicate uncertainty.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ADVICE

Combine Human Annotation With Live Feedback

  • Use both expert annotation and live user feedback to label ambiguous examples; combine gold-standard labels with revealed-preference signals.
  • Sahaj contrasts expensive annotation firms with thumbs-up/down interaction signals used to refine behavior.
INSIGHT

Instruction Tuning Shapes Desired Model Behavior

  • Instruction tuning constrains model outputs by teaching desired behavior with example pairs after pretraining.
  • Sahaj frames instruction tuning as the key post-training step to get models to perform tasks like speech decoding or QA reliably.
ADVICE

Train Tone Preferences With Synthetic Examples And Rewards

  • Create synthetic plus human-labeled examples across tone axes, then fine-tune or RL-train with rewards to match user formality preferences.
  • Sahaj recommends generating variations (formal vs casual) and training reward models to steer outputs.
Get the Snipd Podcast app to discover more snips from this episode
Get the app