SlideShare une entreprise Scribd logo
1  sur  51
Building Tomorrow's Ceph
Sage Weil
Research beginnings
9
UCSC research grant
 “Petascale object storage”
 US Dept of Energy: LANL, LLNL, Sandia
 Scalability
 Reliability
 Performance
 Raw IO bandwidth, metadata ops/sec
 HPC file system workloads
 Thousands of clients writing to same file, directory
Distributed metadata management
 Innovative design
 Subtree-based partitioning for locality, efficiency
 Dynamically adapt to current workload
 Embedded inodes
 Prototype simulator in Java (2004)
 First line of Ceph code
 Summer internship at LLNL
 High security national lab environment
 Could write anything, as long as it was OSS
The rest of Ceph
 RADOS – distributed object storage cluster (2005)
 EBOFS – local object storage (2004/2006)
 CRUSH – hashing for the real world (2005)
 Paxos monitors – cluster consensus (2006)
→ emphasis on consistent, reliable storage
→ scale by pushing intelligence to the edges
→ a different but compelling architecture
 Click to edit the outline text format
 Second Outline Level
Third Outline Level
 Fourth Outline Level
Fifth Outline Level
Sixth Outline Level
Seventh Outline Level
Eighth Outline Level
Ninth Outline LevelClick to edit Master text styles
Industry black hole
 Many large storage vendors
 Proprietary solutions that don't scale well
 Few open source alternatives (2006)
 Very limited scale, or
 Limited community and architecture (Lustre)
 No enterprise feature sets (snapshots, quotas)
 PhD grads all built interesting systems...
 ...and then went to work for Netapp, DDN, EMC, Veritas.
 They want you, not your project
A different path
 Change the world with open source
 Do what Linux did to Solaris, Irix, Ultrix, etc.
 What could go wrong?
 License
 GPL, BSD...
 LGPL: share changes, okay to link to proprietary code
 Avoid community un-friendly practices
 No dual licensing
 No copyright assignment
Incubation
17
DreamHost!
 Move back to Los Angeles, continue hacking
 Hired a few developers
 Pure development
 No deliverables
Ambitious feature set
 Native Linux kernel client (2007-)
 Per-directory snapshots (2008)
 Recursive accounting (2008)
 Object classes (2009)
 librados (2009)
 radosgw (2009)
 strong authentication (2009)
 RBD: rados block device (2010)
The kernel client
 ceph-fuse was limited, not very fast
 Build native Linux kernel implementation
 Began attending Linux file system developer events (LSF)
 Early words of encouragement from ex-Lustre devs
 Engage Linux fs developer community as peer
 Eventually merged CephFS client for v2.6.34 (early 2010)
 RBD client merged in 2011
Part of a larger ecosystem
 Ceph need not solve all problems as monolithic stack
 Replaced ebofs object file system with btrfs
 Same design goals
 Robust, well optimized
 Kernel-level cache management
 Copy-on-write, checksumming, other goodness
 Contributed some early functionality
 Cloning files
 Async snapshots
Budding community
 #ceph on irc.oftc.net, ceph-devel@vger.kernel.org
 Many interested users
 A few developers
 Many fans
 Too unstable for any real deployments
 Still mostly focused on right architecture and technical
solutions
Road to product
 DreamHost decides to build an S3-compatible object
storage service with Ceph
 Stability
 Focus on core RADOS, RBD, radosgw
 Paying back some technical debt
 Build testing automation
 Code review!
 Expand engineering team
The reality
 Growing incoming commercial interest
 Early attempts from organizations large and small
 Difficult to engage with a web hosting company
 No means to support commercial deployments
 Project needed a company to back it
 Fund the engineering effort
 Build and test a product
 Support users
 Bryan built a framework to spin out of DreamHost
Launch
26
Do it right
 How do we build a strong open source company?
 How do we build a strong open source community?
 Models?
 RedHat, Cloudera, MySQL, Canonical, …
 Initial funding from DreamHost, Mark Shuttleworth
Goals
 A stable Ceph release for production deployment
 DreamObjects
 Lay foundation for widespread adoption
 Platform support (Ubuntu, Redhat, SuSE)
 Documentation
 Build and test infrastructure
 Build a sales and support organization
 Expand engineering organization
Branding
 Early decision to engage professional agency
 MetaDesign
 Terms like
 “Brand core”
 “Design system”
 Keep project and company independent
 Inktank != Ceph
 The Future of Storage
 Click to edit the outline text format
Second Outline Level
Slick graphics
 broken powerpoint template
31
Today: adoption
32
Traction
 Too many production deployments to count
 We don't know about most of them!
 Too many customers (for me) to count
 Expansive partner list
 Lots of inbound
 Lots of press and buzz
Quality
 Increased adoption means increased demands on robust
testing
 Across multiple platforms
 Upgrades
 Rolling upgrades
 Inter-version compatibility
Developer community
 Significant external contributors
 Many full-time contributors outside of Inktank
 First-class feature contributions from contributors
 Non-Inktank participants in daily stand-ups
 External access to build/test lab infrastructure
 Common toolset
 Github
 Email (kernel.org)
 IRC (oftc.net)
 Linux distros
CDS: Ceph Developer Summit
 Community process for building project roadmap
 100% online
 Google hangouts
 Wikis
 Etherpad
 Quarterly
 Our 4th
CDS next week
 Great participation
 Ongoing indoctrination of Inktank engineers to open
development model
Erasure coding
 Replication for redundancy is flexible and fast
 For larger clusters, it can be expensive
 Erasure coded data is hard to modify, but ideal for cold or
read-only objects
 Will be used directly by radosgw
 Coexists with new tiering capability
Storage
overhead
Repair
traffic
MTTDL
(days)
3x replication 3x 1x 2.3 E10
RS (10, 4) 1.4x 10x 3.3 E13
LRC (10, 6, 5) 1.6x 5x 1.2 E15
Tiering
 Client side caches are great, but only buy so much.
 Separate hot and cold data onto different storage devices
 Promote hot objects into a faster (e.g., flash-backed) cache pool
 Push cold object back into slower (e.g., erasure-coded) base pool
 Use bloom filters to track temperature
 Common in enterprise solutions; not found in open source
scale-out systems
→ new (with erasure coding) in Firefly release
The Future
40
Technical roadmap
 How do we reach new use-cases and users
 How do we better satisfy existing users
 How do we ensure Ceph can succeed in enough markets
for supporting organizations to thrive
 Enough breadth to expand and grow the community
 Enough focus to do well
Multi-datacenter, geo-replication
 Ceph was originally designed for single DC clusters
 Synchronous replication
 Strong consistency
 Growing demand
 Enterprise: disaster recovery
 ISPs: replication data across sites for locality
 Two strategies:
 use-case specific: radosgw, RBD
 low-level capability in RADOS
RGW: Multi-site and async replication
 Multi-site, multi-cluster
 Regions: east coast, west coast, etc.
 Zones: radosgw sub-cluster(s) within a region
 Can federate across same or multiple Ceph clusters
 Sync user and bucket metadata across regions
 Global bucket/user namespace, like S3
 Synchronize objects across zones
 Within the same region
 Across regions
 Admin control over which zones are master/slave
RBD: block devices
 Today: backup capability
 Based on block device snapshots
 Efficiently mirror changes between consecutive snapshots across
clusters
 Now supported/orchestrated by OpenStack
 Good for coarse synchronization (e.g., hours or days)
 Tomorrow: data journaling for async mirroring
 Pending blueprint at next week's CDS
 Mirror active block device to remote cluster
 Possibly with some configurable delay
Async replication in RADOS
 One implementation to capture multiple use-cases
 RBD, CephFS, RGW, … RADOS
 A harder problem
 Scalable: 1000s OSDs → 1000s of OSDs
 Point-in-time consistency
 Challenging research problem
→ Ongoing design discussion among developers
CephFS
→ This is where it all started – let's get there
 Today
 Stabilization of multi-MDS, directory fragmentation, QA
 NFS, CIFS, Hadoop/HDFS bindings complete but not productized
 Need
 Greater QA investment
 Fsck
 Snapshots
 Amazing community effort (Intel, NUDT and Kylin)
 2014 is the year
Governance
How do we strengthen the project community?
 2014 is the year
 Recognized project leads
 RBD, RGW, RADOS, CephFS, ...
 Formalize emerging processes around CDS, community
roadmap
 External foundation?
The larger ecosystem
The enterprise
How do we pay for all of this?
 Support legacy and transitional client/server interfaces
 iSCSI, NFS, pNFS, CIFS, S3/Swift
 VMWare, Hyper-V
 Identify the beachhead use-cases
 Earn others later
 Single platform – shared storage resource
 Bottom-up: earn respect of engineers and admins
 Top-down: strong brand and compelling product
Why Ceph is the Future of Storage
 It is hard to compete with free and open source software
 Unbeatable value proposition
 Ultimately a more efficient development model
 It is hard to manufacture community
 Strong foundational architecture
 Next-generation protocols, Linux kernel support
 Unencumbered by legacy protocols like NFS
 Move from client/server to client/cluster
 Ongoing paradigm shift
 Software defined infrastructure, data center
 Widespread demand for open platforms
 Click to edit the outline text format
Second Outline Level
Thank you, and Welcome!

Contenu connexe

Tendances

At the Crossroads of HPC and Cloud Computing with Openstack
At the Crossroads of HPC and Cloud Computing with OpenstackAt the Crossroads of HPC and Cloud Computing with Openstack
At the Crossroads of HPC and Cloud Computing with OpenstackRyan Aydelott
 
Deep Dive on REX-Ray, libStorage and the Container Storage Interface - Clinto...
Deep Dive on REX-Ray, libStorage and the Container Storage Interface - Clinto...Deep Dive on REX-Ray, libStorage and the Container Storage Interface - Clinto...
Deep Dive on REX-Ray, libStorage and the Container Storage Interface - Clinto...{code} by Dell EMC
 
Oracle Cloud - Infrastruktura jako kód
Oracle Cloud - Infrastruktura jako kódOracle Cloud - Infrastruktura jako kód
Oracle Cloud - Infrastruktura jako kódMarketingArrowECS_CZ
 
Managing ScaleIO as Software on Mesos
Managing ScaleIO as Software on MesosManaging ScaleIO as Software on Mesos
Managing ScaleIO as Software on MesosDavid vonThenen
 
Why Virtualization is important by Tom Phelan of BlueData
Why Virtualization is important by Tom Phelan of BlueDataWhy Virtualization is important by Tom Phelan of BlueData
Why Virtualization is important by Tom Phelan of BlueDataData Con LA
 
Accelerating Devops via Data Virtualization | Delphix
Accelerating Devops via Data Virtualization | DelphixAccelerating Devops via Data Virtualization | Delphix
Accelerating Devops via Data Virtualization | DelphixDelphixCorp
 
Solving k8s persistent workloads using k8s DevOps style
Solving k8s persistent workloads using k8s DevOps styleSolving k8s persistent workloads using k8s DevOps style
Solving k8s persistent workloads using k8s DevOps styleMayaData
 
OpenEBS Technical Workshop - KubeCon San Diego 2019
OpenEBS Technical Workshop - KubeCon San Diego 2019OpenEBS Technical Workshop - KubeCon San Diego 2019
OpenEBS Technical Workshop - KubeCon San Diego 2019MayaData Inc
 
Container Attached Storage with OpenEBS - CNCF Paris Meetup
Container Attached Storage with OpenEBS - CNCF Paris MeetupContainer Attached Storage with OpenEBS - CNCF Paris Meetup
Container Attached Storage with OpenEBS - CNCF Paris MeetupMayaData Inc
 
Rhs story61712
Rhs story61712Rhs story61712
Rhs story61712rhstorage
 
create auto scale jboss cluster with openshift
create auto scale jboss cluster with openshiftcreate auto scale jboss cluster with openshift
create auto scale jboss cluster with openshiftYusuf Hadiwinata Sutandar
 
Wicked Easy Ceph Block Storage & OpenStack Deployment with Crowbar
Wicked Easy Ceph Block Storage & OpenStack Deployment with CrowbarWicked Easy Ceph Block Storage & OpenStack Deployment with Crowbar
Wicked Easy Ceph Block Storage & OpenStack Deployment with CrowbarCeph Community
 
ttec NAS powered by Open-E
ttec NAS powered by Open-Ettec NAS powered by Open-E
ttec NAS powered by Open-ETTEC
 
Embedded Database Technology | Interbase From Embarcadero Technologies
Embedded Database Technology | Interbase From Embarcadero TechnologiesEmbedded Database Technology | Interbase From Embarcadero Technologies
Embedded Database Technology | Interbase From Embarcadero TechnologiesMichael Findling
 
Make room for more virtual desktops with fast storage
Make room for more virtual desktops with fast storageMake room for more virtual desktops with fast storage
Make room for more virtual desktops with fast storagePrincipled Technologies
 
Optymalizacja środowiska Open Source w celu zwiększenia oszczędności i kontroli
Optymalizacja środowiska Open Source w celu zwiększenia oszczędności i kontroliOptymalizacja środowiska Open Source w celu zwiększenia oszczędności i kontroli
Optymalizacja środowiska Open Source w celu zwiększenia oszczędności i kontroliEDB
 
Hw09 Map Reduce Over Tahoe A Least Authority Encrypted Distributed Filesy...
Hw09   Map Reduce Over Tahoe   A Least Authority Encrypted Distributed Filesy...Hw09   Map Reduce Over Tahoe   A Least Authority Encrypted Distributed Filesy...
Hw09 Map Reduce Over Tahoe A Least Authority Encrypted Distributed Filesy...Cloudera, Inc.
 

Tendances (20)

dbaas-clone
dbaas-clonedbaas-clone
dbaas-clone
 
At the Crossroads of HPC and Cloud Computing with Openstack
At the Crossroads of HPC and Cloud Computing with OpenstackAt the Crossroads of HPC and Cloud Computing with Openstack
At the Crossroads of HPC and Cloud Computing with Openstack
 
Deep Dive on REX-Ray, libStorage and the Container Storage Interface - Clinto...
Deep Dive on REX-Ray, libStorage and the Container Storage Interface - Clinto...Deep Dive on REX-Ray, libStorage and the Container Storage Interface - Clinto...
Deep Dive on REX-Ray, libStorage and the Container Storage Interface - Clinto...
 
Oracle Cloud - Infrastruktura jako kód
Oracle Cloud - Infrastruktura jako kódOracle Cloud - Infrastruktura jako kód
Oracle Cloud - Infrastruktura jako kód
 
Managing ScaleIO as Software on Mesos
Managing ScaleIO as Software on MesosManaging ScaleIO as Software on Mesos
Managing ScaleIO as Software on Mesos
 
Why Virtualization is important by Tom Phelan of BlueData
Why Virtualization is important by Tom Phelan of BlueDataWhy Virtualization is important by Tom Phelan of BlueData
Why Virtualization is important by Tom Phelan of BlueData
 
Accelerating Devops via Data Virtualization | Delphix
Accelerating Devops via Data Virtualization | DelphixAccelerating Devops via Data Virtualization | Delphix
Accelerating Devops via Data Virtualization | Delphix
 
Delphix
DelphixDelphix
Delphix
 
Solving k8s persistent workloads using k8s DevOps style
Solving k8s persistent workloads using k8s DevOps styleSolving k8s persistent workloads using k8s DevOps style
Solving k8s persistent workloads using k8s DevOps style
 
Right Availability in RAC environment. Playing with Oracle clusterware infras...
Right Availability in RAC environment. Playing with Oracle clusterware infras...Right Availability in RAC environment. Playing with Oracle clusterware infras...
Right Availability in RAC environment. Playing with Oracle clusterware infras...
 
OpenEBS Technical Workshop - KubeCon San Diego 2019
OpenEBS Technical Workshop - KubeCon San Diego 2019OpenEBS Technical Workshop - KubeCon San Diego 2019
OpenEBS Technical Workshop - KubeCon San Diego 2019
 
Container Attached Storage with OpenEBS - CNCF Paris Meetup
Container Attached Storage with OpenEBS - CNCF Paris MeetupContainer Attached Storage with OpenEBS - CNCF Paris Meetup
Container Attached Storage with OpenEBS - CNCF Paris Meetup
 
Rhs story61712
Rhs story61712Rhs story61712
Rhs story61712
 
create auto scale jboss cluster with openshift
create auto scale jboss cluster with openshiftcreate auto scale jboss cluster with openshift
create auto scale jboss cluster with openshift
 
Wicked Easy Ceph Block Storage & OpenStack Deployment with Crowbar
Wicked Easy Ceph Block Storage & OpenStack Deployment with CrowbarWicked Easy Ceph Block Storage & OpenStack Deployment with Crowbar
Wicked Easy Ceph Block Storage & OpenStack Deployment with Crowbar
 
ttec NAS powered by Open-E
ttec NAS powered by Open-Ettec NAS powered by Open-E
ttec NAS powered by Open-E
 
Embedded Database Technology | Interbase From Embarcadero Technologies
Embedded Database Technology | Interbase From Embarcadero TechnologiesEmbedded Database Technology | Interbase From Embarcadero Technologies
Embedded Database Technology | Interbase From Embarcadero Technologies
 
Make room for more virtual desktops with fast storage
Make room for more virtual desktops with fast storageMake room for more virtual desktops with fast storage
Make room for more virtual desktops with fast storage
 
Optymalizacja środowiska Open Source w celu zwiększenia oszczędności i kontroli
Optymalizacja środowiska Open Source w celu zwiększenia oszczędności i kontroliOptymalizacja środowiska Open Source w celu zwiększenia oszczędności i kontroli
Optymalizacja środowiska Open Source w celu zwiększenia oszczędności i kontroli
 
Hw09 Map Reduce Over Tahoe A Least Authority Encrypted Distributed Filesy...
Hw09   Map Reduce Over Tahoe   A Least Authority Encrypted Distributed Filesy...Hw09   Map Reduce Over Tahoe   A Least Authority Encrypted Distributed Filesy...
Hw09 Map Reduce Over Tahoe A Least Authority Encrypted Distributed Filesy...
 

En vedette

Reference Architecture: Architecting Ceph Storage Solutions
Reference Architecture: Architecting Ceph Storage Solutions Reference Architecture: Architecting Ceph Storage Solutions
Reference Architecture: Architecting Ceph Storage Solutions Ceph Community
 
London Ceph Day: Ceph Performance and Optimization
London Ceph Day: Ceph Performance and Optimization London Ceph Day: Ceph Performance and Optimization
London Ceph Day: Ceph Performance and Optimization Ceph Community
 
Ceph: Low Fail Go Scale
Ceph: Low Fail Go Scale Ceph: Low Fail Go Scale
Ceph: Low Fail Go Scale Ceph Community
 
Ceph Day Santa Clara: Ceph Fundamentals
Ceph Day Santa Clara: Ceph Fundamentals Ceph Day Santa Clara: Ceph Fundamentals
Ceph Day Santa Clara: Ceph Fundamentals Ceph Community
 
Scaling Ceph at CERN - Ceph Day Frankfurt
Scaling Ceph at CERN - Ceph Day Frankfurt Scaling Ceph at CERN - Ceph Day Frankfurt
Scaling Ceph at CERN - Ceph Day Frankfurt Ceph Community
 
Best Practices with Ceph as Distributed, Intelligent, Unified Cloud Storage -...
Best Practices with Ceph as Distributed, Intelligent, Unified Cloud Storage -...Best Practices with Ceph as Distributed, Intelligent, Unified Cloud Storage -...
Best Practices with Ceph as Distributed, Intelligent, Unified Cloud Storage -...Ceph Community
 
2016-JAN-28 -- High Performance Production Databases on Ceph
2016-JAN-28 -- High Performance Production Databases on Ceph2016-JAN-28 -- High Performance Production Databases on Ceph
2016-JAN-28 -- High Performance Production Databases on CephCeph Community
 
QCT Ceph Solution - Design Consideration and Reference Architecture
QCT Ceph Solution - Design Consideration and Reference ArchitectureQCT Ceph Solution - Design Consideration and Reference Architecture
QCT Ceph Solution - Design Consideration and Reference ArchitectureCeph Community
 
Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons Learned
Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons LearnedCeph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons Learned
Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons LearnedCeph Community
 
CephFS in Jewel: Stable at Last
CephFS in Jewel: Stable at LastCephFS in Jewel: Stable at Last
CephFS in Jewel: Stable at LastCeph Community
 
Ceph Performance and Optimization - Ceph Day Frankfurt
Ceph Performance and Optimization - Ceph Day Frankfurt Ceph Performance and Optimization - Ceph Day Frankfurt
Ceph Performance and Optimization - Ceph Day Frankfurt Ceph Community
 
Ceph Day Taipei - Ceph Tiering with High Performance Architecture
Ceph Day Taipei - Ceph Tiering with High Performance Architecture Ceph Day Taipei - Ceph Tiering with High Performance Architecture
Ceph Day Taipei - Ceph Tiering with High Performance Architecture Ceph Community
 
Ceph Day Tokyo - High Performance Layered Architecture
Ceph Day Tokyo - High Performance Layered Architecture  Ceph Day Tokyo - High Performance Layered Architecture
Ceph Day Tokyo - High Performance Layered Architecture Ceph Community
 
Ceph Day Tokyo - Bit-Isle's 3 years footprint with Ceph
Ceph Day Tokyo - Bit-Isle's 3 years footprint with Ceph Ceph Day Tokyo - Bit-Isle's 3 years footprint with Ceph
Ceph Day Tokyo - Bit-Isle's 3 years footprint with Ceph Ceph Community
 
Ceph Day Tokyo - Delivering cost effective, high performance Ceph cluster
Ceph Day Tokyo - Delivering cost effective, high performance Ceph clusterCeph Day Tokyo - Delivering cost effective, high performance Ceph cluster
Ceph Day Tokyo - Delivering cost effective, high performance Ceph clusterCeph Community
 
Ceph Day Tokyo - Ceph on ARM: Scaleable and Efficient
Ceph Day Tokyo - Ceph on ARM: Scaleable and Efficient Ceph Day Tokyo - Ceph on ARM: Scaleable and Efficient
Ceph Day Tokyo - Ceph on ARM: Scaleable and Efficient Ceph Community
 
Ceph Day Tokyo - Bring Ceph to Enterprise
Ceph Day Tokyo - Bring Ceph to Enterprise Ceph Day Tokyo - Bring Ceph to Enterprise
Ceph Day Tokyo - Bring Ceph to Enterprise Ceph Community
 
Ceph Day Seoul - Ceph: a decade in the making and still going strong
Ceph Day Seoul - Ceph: a decade in the making and still going strong Ceph Day Seoul - Ceph: a decade in the making and still going strong
Ceph Day Seoul - Ceph: a decade in the making and still going strong Ceph Community
 
Ceph Day Tokyo - Ceph Community Update
Ceph Day Tokyo - Ceph Community Update Ceph Day Tokyo - Ceph Community Update
Ceph Day Tokyo - Ceph Community Update Ceph Community
 
Ceph Day London 2014 - Deploying ceph in the wild
Ceph Day London 2014 - Deploying ceph in the wildCeph Day London 2014 - Deploying ceph in the wild
Ceph Day London 2014 - Deploying ceph in the wildCeph Community
 

En vedette (20)

Reference Architecture: Architecting Ceph Storage Solutions
Reference Architecture: Architecting Ceph Storage Solutions Reference Architecture: Architecting Ceph Storage Solutions
Reference Architecture: Architecting Ceph Storage Solutions
 
London Ceph Day: Ceph Performance and Optimization
London Ceph Day: Ceph Performance and Optimization London Ceph Day: Ceph Performance and Optimization
London Ceph Day: Ceph Performance and Optimization
 
Ceph: Low Fail Go Scale
Ceph: Low Fail Go Scale Ceph: Low Fail Go Scale
Ceph: Low Fail Go Scale
 
Ceph Day Santa Clara: Ceph Fundamentals
Ceph Day Santa Clara: Ceph Fundamentals Ceph Day Santa Clara: Ceph Fundamentals
Ceph Day Santa Clara: Ceph Fundamentals
 
Scaling Ceph at CERN - Ceph Day Frankfurt
Scaling Ceph at CERN - Ceph Day Frankfurt Scaling Ceph at CERN - Ceph Day Frankfurt
Scaling Ceph at CERN - Ceph Day Frankfurt
 
Best Practices with Ceph as Distributed, Intelligent, Unified Cloud Storage -...
Best Practices with Ceph as Distributed, Intelligent, Unified Cloud Storage -...Best Practices with Ceph as Distributed, Intelligent, Unified Cloud Storage -...
Best Practices with Ceph as Distributed, Intelligent, Unified Cloud Storage -...
 
2016-JAN-28 -- High Performance Production Databases on Ceph
2016-JAN-28 -- High Performance Production Databases on Ceph2016-JAN-28 -- High Performance Production Databases on Ceph
2016-JAN-28 -- High Performance Production Databases on Ceph
 
QCT Ceph Solution - Design Consideration and Reference Architecture
QCT Ceph Solution - Design Consideration and Reference ArchitectureQCT Ceph Solution - Design Consideration and Reference Architecture
QCT Ceph Solution - Design Consideration and Reference Architecture
 
Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons Learned
Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons LearnedCeph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons Learned
Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons Learned
 
CephFS in Jewel: Stable at Last
CephFS in Jewel: Stable at LastCephFS in Jewel: Stable at Last
CephFS in Jewel: Stable at Last
 
Ceph Performance and Optimization - Ceph Day Frankfurt
Ceph Performance and Optimization - Ceph Day Frankfurt Ceph Performance and Optimization - Ceph Day Frankfurt
Ceph Performance and Optimization - Ceph Day Frankfurt
 
Ceph Day Taipei - Ceph Tiering with High Performance Architecture
Ceph Day Taipei - Ceph Tiering with High Performance Architecture Ceph Day Taipei - Ceph Tiering with High Performance Architecture
Ceph Day Taipei - Ceph Tiering with High Performance Architecture
 
Ceph Day Tokyo - High Performance Layered Architecture
Ceph Day Tokyo - High Performance Layered Architecture  Ceph Day Tokyo - High Performance Layered Architecture
Ceph Day Tokyo - High Performance Layered Architecture
 
Ceph Day Tokyo - Bit-Isle's 3 years footprint with Ceph
Ceph Day Tokyo - Bit-Isle's 3 years footprint with Ceph Ceph Day Tokyo - Bit-Isle's 3 years footprint with Ceph
Ceph Day Tokyo - Bit-Isle's 3 years footprint with Ceph
 
Ceph Day Tokyo - Delivering cost effective, high performance Ceph cluster
Ceph Day Tokyo - Delivering cost effective, high performance Ceph clusterCeph Day Tokyo - Delivering cost effective, high performance Ceph cluster
Ceph Day Tokyo - Delivering cost effective, high performance Ceph cluster
 
Ceph Day Tokyo - Ceph on ARM: Scaleable and Efficient
Ceph Day Tokyo - Ceph on ARM: Scaleable and Efficient Ceph Day Tokyo - Ceph on ARM: Scaleable and Efficient
Ceph Day Tokyo - Ceph on ARM: Scaleable and Efficient
 
Ceph Day Tokyo - Bring Ceph to Enterprise
Ceph Day Tokyo - Bring Ceph to Enterprise Ceph Day Tokyo - Bring Ceph to Enterprise
Ceph Day Tokyo - Bring Ceph to Enterprise
 
Ceph Day Seoul - Ceph: a decade in the making and still going strong
Ceph Day Seoul - Ceph: a decade in the making and still going strong Ceph Day Seoul - Ceph: a decade in the making and still going strong
Ceph Day Seoul - Ceph: a decade in the making and still going strong
 
Ceph Day Tokyo - Ceph Community Update
Ceph Day Tokyo - Ceph Community Update Ceph Day Tokyo - Ceph Community Update
Ceph Day Tokyo - Ceph Community Update
 
Ceph Day London 2014 - Deploying ceph in the wild
Ceph Day London 2014 - Deploying ceph in the wildCeph Day London 2014 - Deploying ceph in the wild
Ceph Day London 2014 - Deploying ceph in the wild
 

Similaire à Keynote: Building Tomorrow's Ceph - Ceph Day Frankfurt

London Ceph Day Keynote: Building Tomorrow's Ceph
London Ceph Day Keynote: Building Tomorrow's Ceph London Ceph Day Keynote: Building Tomorrow's Ceph
London Ceph Day Keynote: Building Tomorrow's Ceph Ceph Community
 
Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph
Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph
Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph Ceph Community
 
Ceph Day NYC: Building Tomorrow's Ceph
Ceph Day NYC: Building Tomorrow's CephCeph Day NYC: Building Tomorrow's Ceph
Ceph Day NYC: Building Tomorrow's CephCeph Community
 
2015 open storage workshop ceph software defined storage
2015 open storage workshop   ceph software defined storage2015 open storage workshop   ceph software defined storage
2015 open storage workshop ceph software defined storageAndrew Underwood
 
The Future of GlusterFS and Gluster.org
The Future of GlusterFS and Gluster.orgThe Future of GlusterFS and Gluster.org
The Future of GlusterFS and Gluster.orgJohn Mark Walker
 
Red Hat Storage 2014 - Product(s) Overview
Red Hat Storage 2014 - Product(s) OverviewRed Hat Storage 2014 - Product(s) Overview
Red Hat Storage 2014 - Product(s) OverviewMarcel Hergaarden
 
Introduction into Ceph storage for OpenStack
Introduction into Ceph storage for OpenStackIntroduction into Ceph storage for OpenStack
Introduction into Ceph storage for OpenStackOpenStack_Online
 
Ceph Day Berlin: Community Update
Ceph Day Berlin:  Community UpdateCeph Day Berlin:  Community Update
Ceph Day Berlin: Community UpdateCeph Community
 
Peanut Butter and jelly: Mapping the deep Integration between Ceph and OpenStack
Peanut Butter and jelly: Mapping the deep Integration between Ceph and OpenStackPeanut Butter and jelly: Mapping the deep Integration between Ceph and OpenStack
Peanut Butter and jelly: Mapping the deep Integration between Ceph and OpenStackSean Cohen
 
Red Hat Storage Day Boston - OpenStack + Ceph Storage
Red Hat Storage Day Boston - OpenStack + Ceph StorageRed Hat Storage Day Boston - OpenStack + Ceph Storage
Red Hat Storage Day Boston - OpenStack + Ceph StorageRed_Hat_Storage
 
UnConference for Georgia Southern Computer Science March 31, 2015
UnConference for Georgia Southern Computer Science March 31, 2015UnConference for Georgia Southern Computer Science March 31, 2015
UnConference for Georgia Southern Computer Science March 31, 2015Christopher Curtin
 
Ceph Day SF 2015 - Keynote
Ceph Day SF 2015 - Keynote Ceph Day SF 2015 - Keynote
Ceph Day SF 2015 - Keynote Ceph Community
 
Ceph Day New York: Ceph: one decade in
Ceph Day New York: Ceph: one decade inCeph Day New York: Ceph: one decade in
Ceph Day New York: Ceph: one decade inCeph Community
 
Bhupeshbansal bigdata
Bhupeshbansal bigdata Bhupeshbansal bigdata
Bhupeshbansal bigdata Bhupesh Bansal
 
A Look at the Future of Storage
A Look at the Future of StorageA Look at the Future of Storage
A Look at the Future of StorageIT Brand Pulse
 
Ceph at salesforce ceph day external presentation
Ceph at salesforce   ceph day external presentationCeph at salesforce   ceph day external presentation
Ceph at salesforce ceph day external presentationSameer Tiwari
 
Webinar - Introduction to Ceph and OpenStack
Webinar - Introduction to Ceph and OpenStackWebinar - Introduction to Ceph and OpenStack
Webinar - Introduction to Ceph and OpenStackCeph Community
 
What you need to know about ceph
What you need to know about cephWhat you need to know about ceph
What you need to know about cephEmma Haruka Iwao
 
Storage and-compute-hdfs-map reduce
Storage and-compute-hdfs-map reduceStorage and-compute-hdfs-map reduce
Storage and-compute-hdfs-map reduceChris Nauroth
 

Similaire à Keynote: Building Tomorrow's Ceph - Ceph Day Frankfurt (20)

London Ceph Day Keynote: Building Tomorrow's Ceph
London Ceph Day Keynote: Building Tomorrow's Ceph London Ceph Day Keynote: Building Tomorrow's Ceph
London Ceph Day Keynote: Building Tomorrow's Ceph
 
Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph
Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph
Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph
 
Ceph Day NYC: Building Tomorrow's Ceph
Ceph Day NYC: Building Tomorrow's CephCeph Day NYC: Building Tomorrow's Ceph
Ceph Day NYC: Building Tomorrow's Ceph
 
2015 open storage workshop ceph software defined storage
2015 open storage workshop   ceph software defined storage2015 open storage workshop   ceph software defined storage
2015 open storage workshop ceph software defined storage
 
The Future of GlusterFS and Gluster.org
The Future of GlusterFS and Gluster.orgThe Future of GlusterFS and Gluster.org
The Future of GlusterFS and Gluster.org
 
Red Hat Storage 2014 - Product(s) Overview
Red Hat Storage 2014 - Product(s) OverviewRed Hat Storage 2014 - Product(s) Overview
Red Hat Storage 2014 - Product(s) Overview
 
Introduction into Ceph storage for OpenStack
Introduction into Ceph storage for OpenStackIntroduction into Ceph storage for OpenStack
Introduction into Ceph storage for OpenStack
 
Ceph as software define storage
Ceph as software define storageCeph as software define storage
Ceph as software define storage
 
Ceph Day Berlin: Community Update
Ceph Day Berlin:  Community UpdateCeph Day Berlin:  Community Update
Ceph Day Berlin: Community Update
 
Peanut Butter and jelly: Mapping the deep Integration between Ceph and OpenStack
Peanut Butter and jelly: Mapping the deep Integration between Ceph and OpenStackPeanut Butter and jelly: Mapping the deep Integration between Ceph and OpenStack
Peanut Butter and jelly: Mapping the deep Integration between Ceph and OpenStack
 
Red Hat Storage Day Boston - OpenStack + Ceph Storage
Red Hat Storage Day Boston - OpenStack + Ceph StorageRed Hat Storage Day Boston - OpenStack + Ceph Storage
Red Hat Storage Day Boston - OpenStack + Ceph Storage
 
UnConference for Georgia Southern Computer Science March 31, 2015
UnConference for Georgia Southern Computer Science March 31, 2015UnConference for Georgia Southern Computer Science March 31, 2015
UnConference for Georgia Southern Computer Science March 31, 2015
 
Ceph Day SF 2015 - Keynote
Ceph Day SF 2015 - Keynote Ceph Day SF 2015 - Keynote
Ceph Day SF 2015 - Keynote
 
Ceph Day New York: Ceph: one decade in
Ceph Day New York: Ceph: one decade inCeph Day New York: Ceph: one decade in
Ceph Day New York: Ceph: one decade in
 
Bhupeshbansal bigdata
Bhupeshbansal bigdata Bhupeshbansal bigdata
Bhupeshbansal bigdata
 
A Look at the Future of Storage
A Look at the Future of StorageA Look at the Future of Storage
A Look at the Future of Storage
 
Ceph at salesforce ceph day external presentation
Ceph at salesforce   ceph day external presentationCeph at salesforce   ceph day external presentation
Ceph at salesforce ceph day external presentation
 
Webinar - Introduction to Ceph and OpenStack
Webinar - Introduction to Ceph and OpenStackWebinar - Introduction to Ceph and OpenStack
Webinar - Introduction to Ceph and OpenStack
 
What you need to know about ceph
What you need to know about cephWhat you need to know about ceph
What you need to know about ceph
 
Storage and-compute-hdfs-map reduce
Storage and-compute-hdfs-map reduceStorage and-compute-hdfs-map reduce
Storage and-compute-hdfs-map reduce
 

Dernier

Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobeapidays
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProduct Anonymous
 
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MIND CTI
 
DBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor PresentationDBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor PresentationDropbox
 
DEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 AmsterdamDEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 AmsterdamUiPathCommunity
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century educationjfdjdjcjdnsjd
 
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Victor Rentea
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoffsammart93
 
Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)Zilliz
 
ICT role in 21st century education and its challenges
ICT role in 21st century education and its challengesICT role in 21st century education and its challenges
ICT role in 21st century education and its challengesrafiqahmad00786416
 
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdfRising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdfOrbitshub
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native ApplicationsWSO2
 
Six Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal OntologySix Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal Ontologyjohnbeverley2021
 
AWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAndrey Devyatkin
 
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Victor Rentea
 
MS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectorsMS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectorsNanddeep Nachan
 
Exploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with MilvusExploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with MilvusZilliz
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FMESafe Software
 
Mcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot Model
Mcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot ModelMcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot Model
Mcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot ModelDeepika Singh
 

Dernier (20)

Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
 
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 
DBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor PresentationDBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor Presentation
 
DEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 AmsterdamDEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
 
Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)
 
ICT role in 21st century education and its challenges
ICT role in 21st century education and its challengesICT role in 21st century education and its challenges
ICT role in 21st century education and its challenges
 
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdfRising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
 
Six Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal OntologySix Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal Ontology
 
AWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of Terraform
 
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
 
MS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectorsMS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectors
 
Exploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with MilvusExploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with Milvus
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
 
Mcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot Model
Mcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot ModelMcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot Model
Mcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot Model
 

Keynote: Building Tomorrow's Ceph - Ceph Day Frankfurt

  • 2.
  • 3.
  • 4.
  • 5.
  • 6.
  • 7.
  • 8.
  • 10.
  • 11. UCSC research grant  “Petascale object storage”  US Dept of Energy: LANL, LLNL, Sandia  Scalability  Reliability  Performance  Raw IO bandwidth, metadata ops/sec  HPC file system workloads  Thousands of clients writing to same file, directory
  • 12. Distributed metadata management  Innovative design  Subtree-based partitioning for locality, efficiency  Dynamically adapt to current workload  Embedded inodes  Prototype simulator in Java (2004)  First line of Ceph code  Summer internship at LLNL  High security national lab environment  Could write anything, as long as it was OSS
  • 13. The rest of Ceph  RADOS – distributed object storage cluster (2005)  EBOFS – local object storage (2004/2006)  CRUSH – hashing for the real world (2005)  Paxos monitors – cluster consensus (2006) → emphasis on consistent, reliable storage → scale by pushing intelligence to the edges → a different but compelling architecture
  • 14.  Click to edit the outline text format  Second Outline Level Third Outline Level  Fourth Outline Level Fifth Outline Level Sixth Outline Level Seventh Outline Level Eighth Outline Level Ninth Outline LevelClick to edit Master text styles
  • 15. Industry black hole  Many large storage vendors  Proprietary solutions that don't scale well  Few open source alternatives (2006)  Very limited scale, or  Limited community and architecture (Lustre)  No enterprise feature sets (snapshots, quotas)  PhD grads all built interesting systems...  ...and then went to work for Netapp, DDN, EMC, Veritas.  They want you, not your project
  • 16. A different path  Change the world with open source  Do what Linux did to Solaris, Irix, Ultrix, etc.  What could go wrong?  License  GPL, BSD...  LGPL: share changes, okay to link to proprietary code  Avoid community un-friendly practices  No dual licensing  No copyright assignment
  • 18.
  • 19. DreamHost!  Move back to Los Angeles, continue hacking  Hired a few developers  Pure development  No deliverables
  • 20. Ambitious feature set  Native Linux kernel client (2007-)  Per-directory snapshots (2008)  Recursive accounting (2008)  Object classes (2009)  librados (2009)  radosgw (2009)  strong authentication (2009)  RBD: rados block device (2010)
  • 21. The kernel client  ceph-fuse was limited, not very fast  Build native Linux kernel implementation  Began attending Linux file system developer events (LSF)  Early words of encouragement from ex-Lustre devs  Engage Linux fs developer community as peer  Eventually merged CephFS client for v2.6.34 (early 2010)  RBD client merged in 2011
  • 22. Part of a larger ecosystem  Ceph need not solve all problems as monolithic stack  Replaced ebofs object file system with btrfs  Same design goals  Robust, well optimized  Kernel-level cache management  Copy-on-write, checksumming, other goodness  Contributed some early functionality  Cloning files  Async snapshots
  • 23. Budding community  #ceph on irc.oftc.net, ceph-devel@vger.kernel.org  Many interested users  A few developers  Many fans  Too unstable for any real deployments  Still mostly focused on right architecture and technical solutions
  • 24. Road to product  DreamHost decides to build an S3-compatible object storage service with Ceph  Stability  Focus on core RADOS, RBD, radosgw  Paying back some technical debt  Build testing automation  Code review!  Expand engineering team
  • 25. The reality  Growing incoming commercial interest  Early attempts from organizations large and small  Difficult to engage with a web hosting company  No means to support commercial deployments  Project needed a company to back it  Fund the engineering effort  Build and test a product  Support users  Bryan built a framework to spin out of DreamHost
  • 27.
  • 28. Do it right  How do we build a strong open source company?  How do we build a strong open source community?  Models?  RedHat, Cloudera, MySQL, Canonical, …  Initial funding from DreamHost, Mark Shuttleworth
  • 29. Goals  A stable Ceph release for production deployment  DreamObjects  Lay foundation for widespread adoption  Platform support (Ubuntu, Redhat, SuSE)  Documentation  Build and test infrastructure  Build a sales and support organization  Expand engineering organization
  • 30. Branding  Early decision to engage professional agency  MetaDesign  Terms like  “Brand core”  “Design system”  Keep project and company independent  Inktank != Ceph  The Future of Storage
  • 31.  Click to edit the outline text format Second Outline Level Slick graphics  broken powerpoint template 31
  • 33.
  • 34. Traction  Too many production deployments to count  We don't know about most of them!  Too many customers (for me) to count  Expansive partner list  Lots of inbound  Lots of press and buzz
  • 35. Quality  Increased adoption means increased demands on robust testing  Across multiple platforms  Upgrades  Rolling upgrades  Inter-version compatibility
  • 36. Developer community  Significant external contributors  Many full-time contributors outside of Inktank  First-class feature contributions from contributors  Non-Inktank participants in daily stand-ups  External access to build/test lab infrastructure  Common toolset  Github  Email (kernel.org)  IRC (oftc.net)  Linux distros
  • 37. CDS: Ceph Developer Summit  Community process for building project roadmap  100% online  Google hangouts  Wikis  Etherpad  Quarterly  Our 4th CDS next week  Great participation  Ongoing indoctrination of Inktank engineers to open development model
  • 38. Erasure coding  Replication for redundancy is flexible and fast  For larger clusters, it can be expensive  Erasure coded data is hard to modify, but ideal for cold or read-only objects  Will be used directly by radosgw  Coexists with new tiering capability Storage overhead Repair traffic MTTDL (days) 3x replication 3x 1x 2.3 E10 RS (10, 4) 1.4x 10x 3.3 E13 LRC (10, 6, 5) 1.6x 5x 1.2 E15
  • 39. Tiering  Client side caches are great, but only buy so much.  Separate hot and cold data onto different storage devices  Promote hot objects into a faster (e.g., flash-backed) cache pool  Push cold object back into slower (e.g., erasure-coded) base pool  Use bloom filters to track temperature  Common in enterprise solutions; not found in open source scale-out systems → new (with erasure coding) in Firefly release
  • 41. Technical roadmap  How do we reach new use-cases and users  How do we better satisfy existing users  How do we ensure Ceph can succeed in enough markets for supporting organizations to thrive  Enough breadth to expand and grow the community  Enough focus to do well
  • 42. Multi-datacenter, geo-replication  Ceph was originally designed for single DC clusters  Synchronous replication  Strong consistency  Growing demand  Enterprise: disaster recovery  ISPs: replication data across sites for locality  Two strategies:  use-case specific: radosgw, RBD  low-level capability in RADOS
  • 43. RGW: Multi-site and async replication  Multi-site, multi-cluster  Regions: east coast, west coast, etc.  Zones: radosgw sub-cluster(s) within a region  Can federate across same or multiple Ceph clusters  Sync user and bucket metadata across regions  Global bucket/user namespace, like S3  Synchronize objects across zones  Within the same region  Across regions  Admin control over which zones are master/slave
  • 44. RBD: block devices  Today: backup capability  Based on block device snapshots  Efficiently mirror changes between consecutive snapshots across clusters  Now supported/orchestrated by OpenStack  Good for coarse synchronization (e.g., hours or days)  Tomorrow: data journaling for async mirroring  Pending blueprint at next week's CDS  Mirror active block device to remote cluster  Possibly with some configurable delay
  • 45. Async replication in RADOS  One implementation to capture multiple use-cases  RBD, CephFS, RGW, … RADOS  A harder problem  Scalable: 1000s OSDs → 1000s of OSDs  Point-in-time consistency  Challenging research problem → Ongoing design discussion among developers
  • 46. CephFS → This is where it all started – let's get there  Today  Stabilization of multi-MDS, directory fragmentation, QA  NFS, CIFS, Hadoop/HDFS bindings complete but not productized  Need  Greater QA investment  Fsck  Snapshots  Amazing community effort (Intel, NUDT and Kylin)  2014 is the year
  • 47. Governance How do we strengthen the project community?  2014 is the year  Recognized project leads  RBD, RGW, RADOS, CephFS, ...  Formalize emerging processes around CDS, community roadmap  External foundation?
  • 49. The enterprise How do we pay for all of this?  Support legacy and transitional client/server interfaces  iSCSI, NFS, pNFS, CIFS, S3/Swift  VMWare, Hyper-V  Identify the beachhead use-cases  Earn others later  Single platform – shared storage resource  Bottom-up: earn respect of engineers and admins  Top-down: strong brand and compelling product
  • 50. Why Ceph is the Future of Storage  It is hard to compete with free and open source software  Unbeatable value proposition  Ultimately a more efficient development model  It is hard to manufacture community  Strong foundational architecture  Next-generation protocols, Linux kernel support  Unencumbered by legacy protocols like NFS  Move from client/server to client/cluster  Ongoing paradigm shift  Software defined infrastructure, data center  Widespread demand for open platforms
  • 51.  Click to edit the outline text format Second Outline Level Thank you, and Welcome!