Data Engineering Podcast

Ceph: A Reliable And Scalable Distributed Filesystem with Sage Weil - Episode 40

Jul 16, 2018
Sage Weil, creator and lead maintainer of Ceph and Red Hat engineer, talks about the origins and architecture of a scalable distributed storage system. He covers object, block, and file interfaces. He explains fault-tolerance, daemon roles, multi-site replication, Kubernetes integration, hardware sizing, and plans for performance and multi-cloud data services.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ANECDOTE

Ceph Began As An Academic Project Turned Production

  • Sage Weil started Ceph as a research project at UC Santa Cruz focused on petabyte-scale HPC storage.
  • After years of development it later became production-ready and was used at DreamHost for an S3-compatible service and block storage around 2011–2012.
INSIGHT

RADOS Provides A Single Resilient Underlayer

  • Ceph's architecture centers on an underlying object layer (RADOS) designed to avoid single points of failure and scale with unreliable commodity hardware.
  • Building block and file interfaces on top of RADOS allowed different optimizations per API while sharing a single resilient storage substrate.
ADVICE

Require Monitor Quorum To Prevent Split Brain

  • Use a monitor quorum to avoid split-brain: Ceph requires a majority of monitors online to form a consistent view and allow writes.
  • This lets the system avoid heavyweight consensus on every IO and instead use Paxos only for cluster membership decisions.
Get the Snipd Podcast app to discover more snips from this episode
Get the app