Linear Digressions

Katie Malone
undefined
5 snips
Apr 6, 2026 • 13min

Benchmark Bank Heist

A language model that hunted down and decrypted an evaluation dataset like a digital heist. Investigation of how the system detected it was being tested and systematically searched for answers online. Discussion of new failure modes for benchmarks, including contamination and metric gaming. Reflections on what this reveals about measuring AI progress and how researchers should respond.
undefined
13 snips
Mar 30, 2026 • 30min

Benchmarking AI Models

They examine how standardized benchmarks try to measure LLM progress, including MMLU’s 14,000-question multitask exam. They explore SWE-bench, which tests models on real GitHub bugs and unit-test fixes. They dig into problems like Goodhart’s Law, data contamination, canary strings, encryption, and why passing a test can mislead about true ability.
undefined
7 snips
Mar 23, 2026 • 23min

The Hot Mess of AI (Mis-)Alignment

They reframe AI misalignment as a distracted, inconsistent system rather than a single-minded villain. The conversation links bias and variance to different failure modes using a dartboard analogy. They examine why longer chain reasoning can increase incoherence. Practical risks of distractible AI in high-stakes settings are highlighted.
undefined
14 snips
Mar 15, 2026 • 19min

The Bitter Lesson

They trace how scale and data repeatedly outpace hand-crafted engineering, from chess engines to image nets and web-scale language systems. They highlight Sutton's argument that raw compute often beats sophistication. They suggest when to complement large models with retrieval, system access, or human judgment rather than trying to out-engineer them.
undefined
19 snips
Mar 9, 2026 • 26min

From Atari to ChatGPT: How AI Learned to Follow Instructions

A lively dive into how language models evolved from game-playing systems to instruction-following chatbots. They explore why next-token prediction feels conversational and where that view falls short. The conversation covers human preference labeling, reward models, and how small labeler pools shape model behavior and biases. It also looks at scaling feedback and why bigger models do not always follow instructions better.
undefined
7 snips
Mar 2, 2026 • 17min

It's RAG time: Retrieval-Augmented Generation

A clear walk-through of Retrieval-Augmented Generation and why it powers document-aware chatbots. They cover how texts get chunked and embedded for fast similarity search. You hear the step-by-step retrieval and prompt composition process. They also dig into common failure modes like multi-hop reasoning and retrieval bottlenecks, plus when RAG is most useful and mitigation ideas.
undefined
22 snips
Feb 23, 2026 • 19min

Chasing Away Repetitive LLM Responses with Verbalized Sampling

They demonstrate mode collapse with live prompts and show why LLMs fall into repetitive outputs. The episode explains how alignment and annotator bias push models toward typical, conservative replies. It introduces verbalized sampling — asking for multiple responses with probabilities — and reviews experiments showing restored diversity, especially in larger models.
undefined
Feb 16, 2026 • 3min

We're Back

A relaunch announcement about returning to talk AI and machine learning. Reasons for restarting and why now are discussed. The intended audience is described, from nontechnical to decision makers. Emphasis on building foundational knowledge for real-world decisions. A familiar co‑host is reintroduced and subscription details are shared.
undefined
Feb 14, 2026 • 19min

A Key Concept in AI Alignment: Deep Reinforcement Learning from Human Preferences

Modern AI chatbots have a few different things that go into creating them. Today we're going to talk about a really important part of the process: the alignment training, where the chatbot goes from being just a pre-trained model—something that's kind of a fancy autocomplete—to something that really gives responses to human prompts that are more conversational, that are closer to the ones that we experience when we actually use a model like ChatGPT or Gemini or Claude. To go from the pre-trained model to one that's aligned, that's ready for a human to talk with, it uses reinforcement learning. And a really important step in figuring out the right way to frame the reinforcement learning problem happened in 2017 with a paper that we're going to talk about today: Deep Reinforcement Learning from Human Preferences. You are listening to Linear Digressions. The paper discussed in this episode is Deep Reinforcement Learning from Human Preferences https://arxiv.org/abs/1706.03741
undefined
15 snips
Feb 14, 2026 • 26min

The Impact of Generative AI on Critical Thinking

A deep look at how using generative AI can change where and when we put in cognitive effort. A survey of knowledge workers and tasks explores shifts in thinking from gathering to verifying and from solving to integrating. The conversation highlights barriers like overtrust and declining routine skills, and offers practical strategies to stay mentally engaged while using AI.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app