
Software Engineering Radio - the podcast for professional software developers SE Radio 715: Sahaj Garg on Designing for Ambiguity in Human Input
5 snips
Apr 8, 2026 Sahaj Garg, co-founder and CTO of Wispr and former AI engineer with ML research, discusses designing for ambiguity in human input. He explores categories like lexical, syntactic, and voice-specific ambiguity. Practical topics include using context and personalization, dataset construction and annotation, inference tradeoffs, revealed preferences from user edits, and subtle ways to communicate uncertainty.
AI Snips
Chapters
Transcript
Episode notes
Combine Human Annotation With Live Feedback
- Use both expert annotation and live user feedback to label ambiguous examples; combine gold-standard labels with revealed-preference signals.
- Sahaj contrasts expensive annotation firms with thumbs-up/down interaction signals used to refine behavior.
Instruction Tuning Shapes Desired Model Behavior
- Instruction tuning constrains model outputs by teaching desired behavior with example pairs after pretraining.
- Sahaj frames instruction tuning as the key post-training step to get models to perform tasks like speech decoding or QA reliably.
Train Tone Preferences With Synthetic Examples And Rewards
- Create synthetic plus human-labeled examples across tone axes, then fine-tune or RL-train with rewards to match user formality preferences.
- Sahaj recommends generating variations (formal vs casual) and training reward models to steer outputs.
