
Eye On A.I. #316 Robbie Goldfarb: Why the Future of AI Depends on Better Judgment
14 snips
Jan 23, 2026 Robbie Goldfarb, former Meta product leader and co-founder of Forum AI, builds systems to scale expert judgment for evaluating AI. He discusses why treating AI as a truth engine is dangerous. Short takes cover expert-driven evaluation, capturing expert thought processes, consequence mapping for mental health and politics, sycophancy and trust, and why transparency about who trains models matters.
AI Snips
Chapters
Books
Transcript
Episode notes
Capture Expert Thought As Graphs
- Experts' thought processes can be captured and represented as graph-like chains to build reliable evaluators for AI.
- Agentic systems that mirror those expert graphs help generalize expert judgment across similar domain questions.
Elicit How Experts Reason, Then Test
- Ask experts to describe how they would answer questions, not just their final answers, to expose decision chains.
- Test and split judges when thought-process patterns fail to generalize across scenarios.
Label Outcomes, Not Just Correctness
- Consequence mapping asks experts what outcomes a model interaction would produce rather than just labeling it good or bad.
- This richer labeling reveals why decisions are harmful or helpful and improves evaluation and prompting strategies.



