
Data Engineering Podcast Ceph: A Reliable And Scalable Distributed Filesystem with Sage Weil - Episode 40
Jul 16, 2018
Sage Weil, creator and lead maintainer of Ceph and Red Hat engineer, talks about the origins and architecture of a scalable distributed storage system. He covers object, block, and file interfaces. He explains fault-tolerance, daemon roles, multi-site replication, Kubernetes integration, hardware sizing, and plans for performance and multi-cloud data services.
AI Snips
Chapters
Transcript
Episode notes
Ceph Began As An Academic Project Turned Production
- Sage Weil started Ceph as a research project at UC Santa Cruz focused on petabyte-scale HPC storage.
- After years of development it later became production-ready and was used at DreamHost for an S3-compatible service and block storage around 2011–2012.
RADOS Provides A Single Resilient Underlayer
- Ceph's architecture centers on an underlying object layer (RADOS) designed to avoid single points of failure and scale with unreliable commodity hardware.
- Building block and file interfaces on top of RADOS allowed different optimizations per API while sharing a single resilient storage substrate.
Require Monitor Quorum To Prevent Split Brain
- Use a monitor quorum to avoid split-brain: Ceph requires a majority of monitors online to form a consistent view and allow writes.
- This lets the system avoid heavyweight consensus on every IO and instead use Paxos only for cluster membership decisions.
