DevOps Paradox

DOP 344: KubeCon EU 2026 Review

Apr 1, 2026
A KubeCon EU 2026 review that spotlights Kubernetes shifting into an AI and inference platform. They cover vendor contributions like NVIDIA and Google, new CNCF sandboxes, and model routing becoming a networking primitive. Expect debates on micro VMs for secure inference, agents as first-class platform users, platform engineering culture bottlenecks, and whether CNCF is absorbing projects companies no longer maintain.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Kubernetes Became An AI Platform

  • KubeCon EU 2026 showed Kubernetes shifting from a container orchestrator to an AI-first platform focused on inference and orchestration of non-container workloads.
  • NVIDIA, Google, and Red Hat contributions (DRA drivers, KAI Scheduler, LLM-D) highlighted infrastructure-level AI integration across GPUs and TPUs.
INSIGHT

Inference Pushes Operators Toward Micro VMs

  • Inference workloads change the game: containers may be insufficient because models can 'escape', pushing operators toward micro VMs and different runtime choices.
  • Victor and Whitney noted startup time and image size matter far less for terabyte-scale model weights than safe isolation and hardware support.
INSIGHT

Model Routing Moves Into Kubernetes Networking

  • Gateway API gained an inference extension, making model routing a first-class Kubernetes networking primitive tied to user sessions and multi-cluster inference.
  • This elevates model routing needs into core networking concerns for large-scale, multi-node inference deployments.
Get the Snipd Podcast app to discover more snips from this episode
Get the app