ThursdAI - The top AI news from the past week

AGI is here? Jensen says yes, ARC-AGI-3 says AI scores under 1%

198 snips
Mar 27, 2026
Daniel Han, open-source ML engineer and Onslaught contributor who builds tools for fine-tuning and quantization, joins to unpack model compression and deployment. He explains weight quantization, KV cache trade-offs, and real-world limits of TurboQuant. The conversation also reacts to Gemini 3.1 Flash Live, new voice and music models, and key infra and quota dramas in the AI world.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ADVICE

Use Subconversations For Fast Follow-Ups While Agents Run

  • Use 'slash btw' or similar sub-conversation features to record follow-ups without triggering tool calls and get fast answers.
  • Alex described OpenClaw/OpenCall's slash btw that spawns a lightweight chat for quick follow-ups while long tasks run.
INSIGHT

Agents Tend To Inflate Code With Irrelevant Defensive Logic

  • Agentic code generation accumulates defensive, irrelevant checks over time, making generated code brittle and bloated.
  • Yam and Mario Zichner's 'Slowing The Fuck Down' visualization showed agents adding layers of checks that degrade long-term code quality.
ANECDOTE

OpenAI Sunsets Sora To Reallocate GPUs Toward Core Models

  • OpenAI shut down Sora app, API, and the $1B Disney deal to refocus compute on core models like Codex and GPT 5.4.
  • Alex explained the reason: GPU costs and focus on AGI-aligned priorities, not moral judgment about content.
Get the Snipd Podcast app to discover more snips from this episode
Get the app