
Software Engineering Daily DeepMind’s RAG System with Animesh Chatterji and Ivan Solovyev
104 snips
Mar 12, 2026 Animesh Chatterji, engineering lead at DeepMind who built production RAG systems, and Ivan Solovyev, product lead for the File Search tool, discuss a managed RAG approach. They talk about simplifying pricing and indexing, advances in embeddings and chunking, default retrieval settings, multimodal retrieval plans, and tradeoffs between configurability and ease of use.
AI Snips
Chapters
Transcript
Episode notes
Map Model Citations To Chunk IDs For Grounding
- Use model-generated citations plus stored chunk IDs to map answers back to original documents.
- Animesh explained the model cites unique indexes for each returned chunk so post-processing can map citations to source metadata.
Beam Uses File Search To Teach New Game Developers
- Beam used File Search to onboard game developers by indexing engine code and documentation.
- Ivan described Beam pulling docs into agent context to guide novices on modules, animations, and scripts during development.
RAG Latency Is Small But Quality Varies By Domain
- Retrieval latency is a few seconds and retrieval accuracy varies by domain, sometimes reaching ~85% correct hits.
- Ivan said retrieval latency aligns with model latency and quality depends on the dataset and use case.



