

Interconnects
Nathan Lambert
Audio essays about the latest developments in AI and interviews with leading scientists in the field. Breaking the hype, understanding what's under the hood, and telling stories. www.interconnects.ai
Episodes
Mentioned books

56 snips
Mar 22, 2026 • 13min
Lossy self-improvement
Debate over whether AI will accelerate itself into a rapid takeoff or hit practical limits. Definitions and history of recursive self-improvement are explored. Technical, political, and economic frictions that slow self-improvement are highlighted. Discussions cover AutoML lessons, diminishing returns from many agents, and why progress may feel linear rather than explosive.

62 snips
Mar 18, 2026 • 7min
GPT 5.4 is a big step for Codex
A lively take on how GPT 5.4 advances agent workflows by improving correctness, speed, ease of use, and cost. Discussion covers everyday engineering tasks that used to cause frequent failures and why the new model feels smoother. Comparisons highlight contrasting styles and practical trade offs between different AI systems. Thoughts on Codex app polish, token efficiency, and future integrations round out the conversation.

46 snips
Mar 16, 2026 • 18min
What comes next with open models
A look at why 2025 pushed many companies to release open AI models and how one breakout win shifted strategies. A discussion of whether open models can economically compete with closed labs and the persistent performance gap. A breakdown of three future model classes and why small, specialized open models may be the most practical opportunity. Thoughts on systems, tools, and building diverse ecosystems instead of chasing frontier scale.

47 snips
Mar 6, 2026 • 36min
Dean Ball on open models and government control
Dean W. Ball, policy and governance commentator and author of the Hyperdimensional newsletter, explores how the Anthropic vs. DoW clash reshapes trust in open models. He discusses open weights as insurance against concentrated control. They cover funding paths, sovereign and regional initiatives, infrastructure and tooling gaps, and why open efforts may win long-term despite short-term hurdles.

28 snips
Mar 5, 2026 • 11min
Olmo Hybrid and future LLM architectures
Discussion of hybrid LLM architectures that mix RNN-like modules with attention. A look back at early hybrids like Mamba and why they struggled when scaled. Exploration of Gated DeltaNet versus Mamba approaches and where Olmo Hybrid sits. Coverage of scaling studies, layer-ratio effects, pretraining efficiency gains, and practical tooling and stability tradeoffs affecting throughput.

46 snips
Feb 24, 2026 • 11min
How much does distillation really matter for Chinese LLMs?
The conversation unpacks modern distillation as synthetic data from stronger models and why it powers day-to-day model improvement. It surveys allegations that Chinese labs used API outputs to reproduce capabilities and examines specific cases and token volumes. It contrasts distillation’s practical limits with large-scale reinforcement learning and explores why resource constraints push labs toward synthetic-data shortcuts.

111 snips
Feb 9, 2026 • 8min
Opus 4.6, Codex 5.3, and the post-benchmark era
A deep comparison of the latest coding models from OpenAI and Anthropic. They discuss changes in agent workflows and real-world usability. Tradeoffs surface between speed, ease of use, and reliability. Practical differences in bug finding, product polish, and who each model serves are highlighted.

32 snips
Feb 4, 2026 • 1h 8min
Why Nvidia builds open models with Bryan Catanzaro
Bryan Catanzaro, VP of Applied Deep Learning Research at NVIDIA and leader of the Nemotron effort, discusses Nemotron models and why NVIDIA releases high-quality open models. He covers model and dataset releases, team culture and coordination, Megatron-LM and training software, engineering work to make models usable, and how open models fit NVIDIA’s business and systems strategy.

74 snips
Jan 30, 2026 • 11min
Thoughts on the job market in the age of LLMs
A candid look at hiring frictions in AI, from fierce competition for senior talent to barriers facing juniors. Discussion of why senior roles rise in importance as agents handle routine work. Tales of how obsessive drive, visible work, and targeted outreach win hires. Reflections on short-term hiring rhythms, open-source limits, and choosing jobs for visibility and responsibility.

29 snips
Jan 27, 2026 • 1h 12min
Arcee AI goes all-in on open models built in the U.S.
Lucas Atkins, CTO and pretraining/architecture lead at Arcee AI, talks about building Trinity Large and the six-month sprint to train a 400B total, 13B active MoE. He covers model scaling, Muon optimizer choices, stability and expert balancing, dataset and compute planning on B300 hardware, and deployment, licensing, and business tradeoffs for open U.S.-built models.


