SlideShare a Scribd company logo
1 of 40
Download to read offline
COMMUNITY UPDATE
SAGE WEIL – RED HAT
2017.05.11
2
UPSTREAM RELEASES
Jewel (LTS)
Spring 2016
Kraken
Fall 2016
Luminous (LTS)
Spring 2017
12.2.z
10.2.z
14.2.z
WE ARE HERE
Mimic
Fall 2017?
Nautilus? (LTS)
Spring 2018?
3 LUMINOUS
4
BLUESTORE: STABLE AND DEFAULT
● New OSD backend
– consumes raw block device(s) – no more XFS
– embeds rocksdb for metadata
● Fast on both HDDs (~2x) and SSDs (~1.5x)
– Similar to FileStore on NVMe, where the device is not the bottleneck
● Smaller journals
– happily uses fast SSD partition(s) for internal metadata, or NVRAM for journal
● Full data checksums (crc32c, xxhash, etc.)
● Inline compression (zlib, snappy)
– policy driven by global or per-pool config, and/or client hints
● Stable and default
5
HDD: RANDOM WRITE
4 8 16 32 64
128
256
512
1024
2048
4096
0
100
200
300
400
500
Bluestore vs Filestore HDD Random Write Throughput
Filestore
Bluestore (wip-bitmap-alloc-perf)
BS (Master-a07452d)
BS (Master-d62a4948)
Bluestore (wip-bluestore-dw)
IO Size
Throughput(MB/s)
4 8 16 32 64
128
256
512
1024
2048
4096
0
500
1000
1500
2000
Bluestore vs Filestore HDD Random Write IOPS
Filestore
Bluestore (wip-bitmap-alloc-perf)
BS (Master-a07452d)
BS (Master-d62a4948)
Bluestore (wip-bluestore-dw)
IO Size
IOPS
6
HDD: MIXED READ/WRITE
4 8 16 32 64
128
256
512
1024
2048
4096
0
50
100
150
200
250
300
350
Bluestore vs Filestore HDD Random RWThroughput
Filestore
Bluestore (wip-bitmap-alloc-perf)
BS (Master-a07452d)
BS (Master-d62a4948)
Bluestore (wip-bluestore-dw)
IO Size
Throughput(MB/s)
4 8 16 32 64
128
256
512
1024
2048
4096
0
200
400
600
800
1000
1200
Bluestore vs Filestore HDD Random RW IOPS
Filestore
Bluestore (wip-bitmap-alloc-perf)
BS (Master-a07452d)
BS (Master-d62a4948)
Bluestore (wip-bluestore-dw)
IO Size
IOPS
7
RGW ON HDD+NVME, EC 4+2
1 Bucket 128 Buckets 4 Buckets 512 Buckets Rados
1 RGW Server 4 RGW Servers Bench
0
200
400
600
800
1000
1200
1400
1600
1800
4+2 Erasure Coding RadosGW Write Tests
32MB Objects, 24 HDD/NVMe OSDs on 4 Servers, 4 Clients
Filestore 512KB Chunks
Filestore 4MB Chunks
Bluestore 512KB Chunks
Bluestore 4MB Chunks
Throughput(MB/s)
8
RBD OVER ERASURE CODED POOLS
● aka erasure code overwrites
● requires BlueStore to perform reasonably
● significant improvement in efficiency over 3x replication
– 2+2 → 2x 4+2 → 1.5x
● small writes slower than replication
– early testing showed 4+2 is about half as fast as 3x replication
● large writes faster than replication
– less IO to device
● implementation still does the “simple” thing
– all writes update a full stripe
9
CEPH-MGR
● ceph-mgr
– new management daemon to supplement ceph-mon (monitor)
– easier integration point for python management logic
– integrated metrics
● make ceph-mon scalable again
– offload pg stats from mon to mgr
– push to 10K OSDs (planned “big bang 3” @ CERN)
● new REST API
– pecan
– based on previous Calamari API
● built-in web dashboard
– webby equivalent of 'ceph -s'
M G
???
(time for new iconography)
10
11
ASYNCMESSENGER
●
new network Messenger implementation
– event driven
– fixed-size thread pool
●
RDMA backend (ibverbs)
– built by default
– limited testing, but seems stable!
●
DPDK backend
– prototype!
12
PERFECTLY BALANCED OSDS (FINALLY)
● CRUSH choose_args
– alternate weight sets for individual rules
– complete flexibility to optimize weights etc
– fixes two problems
● imbalance – run numeric optimization to adjust weights to balance PG
distribution for a pool (or cluster)
● multipick anomaly – adjust weights per position to correct for low-
weighted devices (e.g., mostly empty rack)
– backward compatible with pre-luminous clients for imbalance case
● pg upmap
– explicitly map individual PGs to specific devices in OSDMap
– simple offline optimizer balance PGs
– by pool or by cluster
– requires luminous+ clients
13
RADOS MISC
●
CRUSH device classes
– mark OSDs with class (hdd, ssd, etc)
– out-of-box rules to map to specific class of devices within the
same hierarchy
●
streamlined disk replacement
●
require_min_compat_client – simpler, safer configuration
●
annotated/documented config options
●
client backoff on stuck PGs or objects
●
better EIO handling
●
peering and recovery speedups
●
fast OSD failure detection
14
S3
Swift
Erasure coding
Multisite federation
Multisite replication
NFS
Encryption
Tiering
Deduplication
RADOSGW
Compression
15
ZONE CZONE B
RGW METADATA SEARCH
RADOSGW
LIBRADOS
M
CLUSTER A
MM M
CLUSTER B
MM
RADOSGW RADOSGW RADOSGW
LIBRADOS LIBRADOS LIBRADOS
REST
RADOSGW
LIBRADOS
M
CLUSTER C
MM
REST
ZONE A
16
RGW MISC
●
NFS gateway
– NFSv4 and v3
– full object access (not general purpose!)
●
dynamic bucket index sharding
– automatic (finally!)
●
inline compression
●
encryption
– follows S3 encryption APIs
●
S3 and Swift API odds and ends
RADOSGW
LIBRADOS
17
Erasure coding
Multisite mirroring
Persistent client cache
Consistency groups
Encryption
iSCSI
Trash
RBD
18
RBD
●
RBD over erasure coded pool
– rbd create --data-pool <ecpoolname> ...
●
RBD mirroring improvements
– cooperative HA daemons
– improved Cinder integration
●
iSCSI
– LIO tcmu-runner, librbd (full feature set)
●
Kernel RBD improvements
– exclusive locking, object map
RADOS
A reliable, autonomous, distributed object store comprised of self-healing, self-managing,
intelligent storage nodes
RADOS
A reliable, autonomous, distributed object store comprised of self-healing, self-managing,
intelligent storage nodes
LIBRADOS
A library allowing
apps to directly
access RADOS,
with support for
C, C++, Java,
Python, Ruby,
and PHP
LIBRADOS
A library allowing
apps to directly
access RADOS,
with support for
C, C++, Java,
Python, Ruby,
and PHP
RBD
A reliable and fully-
distributed block
device, with a Linux
kernel client and a
QEMU/KVM driver
RBD
A reliable and fully-
distributed block
device, with a Linux
kernel client and a
QEMU/KVM driver
RADOSGW
A bucket-based
REST gateway,
compatible with S3
and Swift
RADOSGW
A bucket-based
REST gateway,
compatible with S3
and Swift
APPAPP APPAPP HOST/VMHOST/VM CLIENTCLIENT
CEPH FS
A POSIX-compliant
distributed file system,
with a Linux kernel
client and support for
FUSE
CEPH FS
A POSIX-compliant
distributed file system,
with a Linux kernel
client and support for
FUSE
NEARLY
AWESOME
AWESOMEAWESOME
AWESOME
AWESOME
20
2017 =
RGW
A web services gateway
for object storage,
compatible with S3 and
Swift
LIBRADOS
A library allowing apps to directly access RADOS (C, C++, Java, Python, Ruby, PHP)
RADOS
A software-based, reliable, autonomous, distributed object store comprised of
self-healing, self-managing, intelligent storage nodes and lightweight monitors
RBD
A reliable, fully-distributed
block device with cloud
platform integration
CEPHFS
A distributed file system
with POSIX semantics and
scale-out metadata
management
OBJECT BLOCK FILE
FULLY AWESOME
21
CEPHFS
● multiple active MDS daemons (finally!)
● subtree pinning to specific daemon
● directory fragmentation on by default
– (snapshots still off by default)
● so many tests
● so many bugs fixed
● kernel client improvements
22
OSD refactor
Tiering
Client-side caches
Metrics
Dedup
QoS
Self-management
AFTER LUMINOUS?
Multi-site federation
23 MIMIC
24
RADOS
● Peering optimizations
● IO path refactor and optimization
– async, state-driven, futures
– painful but necessary
● BlueStore and rocksdb optimization
– rocksdb level0 compaction
– alternative KV store?
● Erasure coding plugin API improvements
– new codes with less IO for single-OSD failures
DEVICE
OSD
OBJECTSTORE
MESSENGER
25
QUALITY OF SERVICE
● Ongoing background development
– dmclock distributed QoS queuing
– minimum reservations and priority weighting
● Range of policies
– IO type (background, client)
– pool-based
– client-based
● Theory is complex
● Prototype is promising, despite simplicity
● Missing management framework
26
27
CEPH STORAGE CLUSTER
TIERING
● new RADOS ‘redirect’ primitive
– basically a symlink, transparent to librados
– replace “sparse” cache tier with base pool “index”
APPLICATION
BASE POOL (REPLICATED, SSD)
APPLICATION
CACHE POOL (REPLICATED)
BASE POOL (HDD AND/OR ERASURE)
CEPH STORAGE CLUSTER
SLOW 1 (EC) SLOW #1 (...)
28
CEPH STORAGE CLUSTER
DEDUPLICATION WIP
● Generalize redirect to a “manifest”
– map of offsets to object “fragments” (vs a full
object copy)
● Break objects into chunks
– fixed size, or content fingerprint
● Store chunks in content-addressable pool
– name object by sha256(content)
– reference count chunks
● TBD
– inline or post?
– policy (when to promote inline, etc.)
– agent home (embed in OSD, …)
APPLICATION
BASE POOL (REPLICATED, SSD)
SLOW 1 (EC) CAS DEDUP POOL
29
CEPH-MGR: METRICS, MGMT
● Metrics aggregation
– short-term time series out of the box
– no persistent state
– streaming to external platform
(Prometheus, …)
● Host self-management functions
– automatic CRUSH optimization
– identification of slow devices
– steer IO away from busy devices
– device failure prediction
M G
consistency metrics
summary
30
ARM
●
aarch64 builds
– centos7, ubuntu xenial
– have some, but awaiting more build hardware in community lab
●
thank you to partners!
● ppc64
●
armv7l
– http://ceph.com/community/500-osd-ceph-cluster/
31
Gen 2 Gen3
ARM 32bit ARM 64 bit
uBoot UEFI
1GB DDR Up to 4GB DDR
2x 1GbE SGMII 2x 2.5GbE SGMII
Basic Security Enhanced Security
No Flash option Flash option
I2C Management Integrated
Management
Planning Gen3 availability to PoC partners
Contact james.wilshire@wdc.com
WDLABS MICROSERVER UPDATE
32
CLIENT CACHES!
● RGW
– persistent read-only cache on NVMe
– fully consistent (only caches immutable “tail” rados objects)
– Mass Open Cloud
● RBD
– persistent read-only cache of immutable clone parent images
– writeback cache for improving write latency
● cluster image remains crash-consistent if client cache is lost
● CephFS
– kernel client already uses kernel fscache facility
33
GROWING DEVELOPER COMMUNITY
34
● Red Hat
● Mirantis
● SUSE
● ZTE
● China Mobile
● XSky
● Digiware
● Intel
● Kylin Cloud
● Easystack
● Istuary Innovation Group
● Quantum
● Mellanox
● H3C
● Quantum
● UnitedStack
● Deutsche Telekom
● Reliance Jio Infocomm
● OVH
● Alibaba
● DreamHost
● CERN
GROWING DEVELOPER COMMUNITY
35
● Red Hat
● Mirantis
● SUSE
● ZTE
● China Mobile
● XSky
● Digiware
● Intel
● Kylin Cloud
● Easystack
● Istuary Innovation Group
● Quantum
● Mellanox
● H3C
● Quantum
● UnitedStack
● Deutsche Telekom
● Reliance Jio Infocomm
● OVH
● Alibaba
● DreamHost
● CERN
OPENSTACK VENDORS
36
● Red Hat
● Mirantis
● SUSE
● ZTE
● China Mobile
● XSky
● Digiware
● Intel
● Kylin Cloud
● Easystack
● Istuary Innovation Group
● Quantum
● Mellanox
● H3C
● Quantum
● UnitedStack
● Deutsche Telekom
● Reliance Jio Infocomm
● OVH
● Alibaba
● DreamHost
● CERN
CLOUD OPERATORS
37
● Red Hat
● Mirantis
● SUSE
● ZTE
● China Mobile
● XSky
● Digiware
● Intel
● Kylin Cloud
● Easystack
● Istuary Innovation Group
● Quantum
● Mellanox
● H3C
● Quantum
● UnitedStack
● Deutsche Telekom
● Reliance Jio Infocomm
● OVH
● Alibaba
● DreamHost
● CERN
HARDWARE AND SOLUTION VENDORS
38
● Red Hat
● Mirantis
● SUSE
● ZTE
● China Mobile
● XSky
● Digiware
● Intel
● Kylin Cloud
● Easystack
● Istuary Innovation Group
● Quantum
● Mellanox
● H3C
● Quantum
● UnitedStack
● Deutsche Telekom
● Reliance Jio Infocomm
● OVH
● Alibaba
● DreamHost
● CERN
???
39
● Mailing list and IRC
– http://ceph.com/IRC
● Ceph Developer Monthly
– first Weds of every month
– video conference (Bluejeans)
– alternating APAC- and EMEA-
friendly times
● Github
– https://github.com/ceph/
●
Ceph Days
– http://ceph.com/cephdays/
●
Meetups
– http://ceph.com/meetups
●
Ceph Tech Talks
– http://ceph.com/ceph-tech-talks
/
●
‘Ceph’ Youtube channel
– (google it)
●
Twitter
– @ceph
GET INVOLVED
THANK YOU!
Sage Weil
CEPH PRINCIPAL ARCHITECT
sage@redhat.com
@liewegas

More Related Content

What's hot

The State of Ceph, Manila, and Containers in OpenStack
The State of Ceph, Manila, and Containers in OpenStackThe State of Ceph, Manila, and Containers in OpenStack
The State of Ceph, Manila, and Containers in OpenStackSage Weil
 
Making distributed storage easy: usability in Ceph Luminous and beyond
Making distributed storage easy: usability in Ceph Luminous and beyondMaking distributed storage easy: usability in Ceph Luminous and beyond
Making distributed storage easy: usability in Ceph Luminous and beyondSage Weil
 
Keeping OpenStack storage trendy with Ceph and containers
Keeping OpenStack storage trendy with Ceph and containersKeeping OpenStack storage trendy with Ceph and containers
Keeping OpenStack storage trendy with Ceph and containersSage Weil
 
A crash course in CRUSH
A crash course in CRUSHA crash course in CRUSH
A crash course in CRUSHSage Weil
 
Ceph Performance: Projects Leading up to Jewel
Ceph Performance: Projects Leading up to JewelCeph Performance: Projects Leading up to Jewel
Ceph Performance: Projects Leading up to JewelColleen Corrice
 
Ceph at Work in Bloomberg: Object Store, RBD and OpenStack
Ceph at Work in Bloomberg: Object Store, RBD and OpenStackCeph at Work in Bloomberg: Object Store, RBD and OpenStack
Ceph at Work in Bloomberg: Object Store, RBD and OpenStackRed_Hat_Storage
 
CephFS update February 2016
CephFS update February 2016CephFS update February 2016
CephFS update February 2016John Spray
 
Ceph - A distributed storage system
Ceph - A distributed storage systemCeph - A distributed storage system
Ceph - A distributed storage systemItalo Santos
 
Ceph Intro and Architectural Overview by Ross Turk
Ceph Intro and Architectural Overview by Ross TurkCeph Intro and Architectural Overview by Ross Turk
Ceph Intro and Architectural Overview by Ross Turkbuildacloud
 
BlueStore, A New Storage Backend for Ceph, One Year In
BlueStore, A New Storage Backend for Ceph, One Year InBlueStore, A New Storage Backend for Ceph, One Year In
BlueStore, A New Storage Backend for Ceph, One Year InSage Weil
 
Ceph data services in a multi- and hybrid cloud world
Ceph data services in a multi- and hybrid cloud worldCeph data services in a multi- and hybrid cloud world
Ceph data services in a multi- and hybrid cloud worldSage Weil
 
XSKY - ceph luminous update
XSKY - ceph luminous updateXSKY - ceph luminous update
XSKY - ceph luminous updateinwin stack
 
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...Odinot Stanislas
 
Cephfs jewel mds performance benchmark
Cephfs jewel mds performance benchmarkCephfs jewel mds performance benchmark
Cephfs jewel mds performance benchmarkXiaoxi Chen
 
2019.06.27 Intro to Ceph
2019.06.27 Intro to Ceph2019.06.27 Intro to Ceph
2019.06.27 Intro to CephCeph Community
 
Linux Stammtisch Munich: Ceph - Overview, Experiences and Outlook
Linux Stammtisch Munich: Ceph - Overview, Experiences and OutlookLinux Stammtisch Munich: Ceph - Overview, Experiences and Outlook
Linux Stammtisch Munich: Ceph - Overview, Experiences and OutlookDanny Al-Gaaf
 

What's hot (19)

The State of Ceph, Manila, and Containers in OpenStack
The State of Ceph, Manila, and Containers in OpenStackThe State of Ceph, Manila, and Containers in OpenStack
The State of Ceph, Manila, and Containers in OpenStack
 
Making distributed storage easy: usability in Ceph Luminous and beyond
Making distributed storage easy: usability in Ceph Luminous and beyondMaking distributed storage easy: usability in Ceph Luminous and beyond
Making distributed storage easy: usability in Ceph Luminous and beyond
 
Keeping OpenStack storage trendy with Ceph and containers
Keeping OpenStack storage trendy with Ceph and containersKeeping OpenStack storage trendy with Ceph and containers
Keeping OpenStack storage trendy with Ceph and containers
 
A crash course in CRUSH
A crash course in CRUSHA crash course in CRUSH
A crash course in CRUSH
 
Ceph Performance: Projects Leading up to Jewel
Ceph Performance: Projects Leading up to JewelCeph Performance: Projects Leading up to Jewel
Ceph Performance: Projects Leading up to Jewel
 
Ceph at Work in Bloomberg: Object Store, RBD and OpenStack
Ceph at Work in Bloomberg: Object Store, RBD and OpenStackCeph at Work in Bloomberg: Object Store, RBD and OpenStack
Ceph at Work in Bloomberg: Object Store, RBD and OpenStack
 
CephFS update February 2016
CephFS update February 2016CephFS update February 2016
CephFS update February 2016
 
Block Storage For VMs With Ceph
Block Storage For VMs With CephBlock Storage For VMs With Ceph
Block Storage For VMs With Ceph
 
Ceph - A distributed storage system
Ceph - A distributed storage systemCeph - A distributed storage system
Ceph - A distributed storage system
 
Ceph Intro and Architectural Overview by Ross Turk
Ceph Intro and Architectural Overview by Ross TurkCeph Intro and Architectural Overview by Ross Turk
Ceph Intro and Architectural Overview by Ross Turk
 
BlueStore, A New Storage Backend for Ceph, One Year In
BlueStore, A New Storage Backend for Ceph, One Year InBlueStore, A New Storage Backend for Ceph, One Year In
BlueStore, A New Storage Backend for Ceph, One Year In
 
Ceph data services in a multi- and hybrid cloud world
Ceph data services in a multi- and hybrid cloud worldCeph data services in a multi- and hybrid cloud world
Ceph data services in a multi- and hybrid cloud world
 
XSKY - ceph luminous update
XSKY - ceph luminous updateXSKY - ceph luminous update
XSKY - ceph luminous update
 
librados
libradoslibrados
librados
 
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...
 
Cephfs jewel mds performance benchmark
Cephfs jewel mds performance benchmarkCephfs jewel mds performance benchmark
Cephfs jewel mds performance benchmark
 
2019.06.27 Intro to Ceph
2019.06.27 Intro to Ceph2019.06.27 Intro to Ceph
2019.06.27 Intro to Ceph
 
ceph-barcelona-v-1.2
ceph-barcelona-v-1.2ceph-barcelona-v-1.2
ceph-barcelona-v-1.2
 
Linux Stammtisch Munich: Ceph - Overview, Experiences and Outlook
Linux Stammtisch Munich: Ceph - Overview, Experiences and OutlookLinux Stammtisch Munich: Ceph - Overview, Experiences and Outlook
Linux Stammtisch Munich: Ceph - Overview, Experiences and Outlook
 

Similar to Community Update at OpenStack Summit Boston

20171101 taco scargo luminous is out, what's in it for you
20171101 taco scargo   luminous is out, what's in it for you20171101 taco scargo   luminous is out, what's in it for you
20171101 taco scargo luminous is out, what's in it for youTaco Scargo
 
Storage tiering and erasure coding in Ceph (SCaLE13x)
Storage tiering and erasure coding in Ceph (SCaLE13x)Storage tiering and erasure coding in Ceph (SCaLE13x)
Storage tiering and erasure coding in Ceph (SCaLE13x)Sage Weil
 
2021.02 new in Ceph Pacific Dashboard
2021.02 new in Ceph Pacific Dashboard2021.02 new in Ceph Pacific Dashboard
2021.02 new in Ceph Pacific DashboardCeph Community
 
CephFS in Jewel: Stable at Last
CephFS in Jewel: Stable at LastCephFS in Jewel: Stable at Last
CephFS in Jewel: Stable at LastCeph Community
 
Quick-and-Easy Deployment of a Ceph Storage Cluster
Quick-and-Easy Deployment of a Ceph Storage ClusterQuick-and-Easy Deployment of a Ceph Storage Cluster
Quick-and-Easy Deployment of a Ceph Storage ClusterPatrick Quairoli
 
CEPH DAY BERLIN - WHAT'S NEW IN CEPH
CEPH DAY BERLIN - WHAT'S NEW IN CEPH CEPH DAY BERLIN - WHAT'S NEW IN CEPH
CEPH DAY BERLIN - WHAT'S NEW IN CEPH Ceph Community
 
Ceph Block Devices: A Deep Dive
Ceph Block Devices: A Deep DiveCeph Block Devices: A Deep Dive
Ceph Block Devices: A Deep Divejoshdurgin
 
Ceph Block Devices: A Deep Dive
Ceph Block Devices:  A Deep DiveCeph Block Devices:  A Deep Dive
Ceph Block Devices: A Deep DiveRed_Hat_Storage
 
7. Cloud Native Computing - Kubernetes - Bratislava - Rook.io
7. Cloud Native Computing - Kubernetes - Bratislava - Rook.io7. Cloud Native Computing - Kubernetes - Bratislava - Rook.io
7. Cloud Native Computing - Kubernetes - Bratislava - Rook.ioDávid Kőszeghy
 
OSDC 2015: John Spray | The Ceph Storage System
OSDC 2015: John Spray | The Ceph Storage SystemOSDC 2015: John Spray | The Ceph Storage System
OSDC 2015: John Spray | The Ceph Storage SystemNETWAYS
 
Crimson: Ceph for the Age of NVMe and Persistent Memory
Crimson: Ceph for the Age of NVMe and Persistent MemoryCrimson: Ceph for the Age of NVMe and Persistent Memory
Crimson: Ceph for the Age of NVMe and Persistent MemoryScyllaDB
 
Red Hat Gluster Storage Performance
Red Hat Gluster Storage PerformanceRed Hat Gluster Storage Performance
Red Hat Gluster Storage PerformanceRed_Hat_Storage
 
The Hive Think Tank: Ceph + RocksDB by Sage Weil, Red Hat.
The Hive Think Tank: Ceph + RocksDB by Sage Weil, Red Hat.The Hive Think Tank: Ceph + RocksDB by Sage Weil, Red Hat.
The Hive Think Tank: Ceph + RocksDB by Sage Weil, Red Hat.The Hive
 
RBD: What will the future bring? - Jason Dillaman
RBD: What will the future bring? - Jason DillamanRBD: What will the future bring? - Jason Dillaman
RBD: What will the future bring? - Jason DillamanCeph Community
 
Open Source Storage at Scale: Ceph @ GRNET
Open Source Storage at Scale: Ceph @ GRNETOpen Source Storage at Scale: Ceph @ GRNET
Open Source Storage at Scale: Ceph @ GRNETNikos Kormpakis
 

Similar to Community Update at OpenStack Summit Boston (20)

20171101 taco scargo luminous is out, what's in it for you
20171101 taco scargo   luminous is out, what's in it for you20171101 taco scargo   luminous is out, what's in it for you
20171101 taco scargo luminous is out, what's in it for you
 
Storage tiering and erasure coding in Ceph (SCaLE13x)
Storage tiering and erasure coding in Ceph (SCaLE13x)Storage tiering and erasure coding in Ceph (SCaLE13x)
Storage tiering and erasure coding in Ceph (SCaLE13x)
 
2021.02 new in Ceph Pacific Dashboard
2021.02 new in Ceph Pacific Dashboard2021.02 new in Ceph Pacific Dashboard
2021.02 new in Ceph Pacific Dashboard
 
CephFS in Jewel: Stable at Last
CephFS in Jewel: Stable at LastCephFS in Jewel: Stable at Last
CephFS in Jewel: Stable at Last
 
Quick-and-Easy Deployment of a Ceph Storage Cluster
Quick-and-Easy Deployment of a Ceph Storage ClusterQuick-and-Easy Deployment of a Ceph Storage Cluster
Quick-and-Easy Deployment of a Ceph Storage Cluster
 
CEPH DAY BERLIN - WHAT'S NEW IN CEPH
CEPH DAY BERLIN - WHAT'S NEW IN CEPH CEPH DAY BERLIN - WHAT'S NEW IN CEPH
CEPH DAY BERLIN - WHAT'S NEW IN CEPH
 
Ceph Block Devices: A Deep Dive
Ceph Block Devices: A Deep DiveCeph Block Devices: A Deep Dive
Ceph Block Devices: A Deep Dive
 
Ceph Block Devices: A Deep Dive
Ceph Block Devices:  A Deep DiveCeph Block Devices:  A Deep Dive
Ceph Block Devices: A Deep Dive
 
Cncf meetup-rook
Cncf meetup-rookCncf meetup-rook
Cncf meetup-rook
 
Cncf meetup-rook
Cncf meetup-rookCncf meetup-rook
Cncf meetup-rook
 
7. Cloud Native Computing - Kubernetes - Bratislava - Rook.io
7. Cloud Native Computing - Kubernetes - Bratislava - Rook.io7. Cloud Native Computing - Kubernetes - Bratislava - Rook.io
7. Cloud Native Computing - Kubernetes - Bratislava - Rook.io
 
OSDC 2015: John Spray | The Ceph Storage System
OSDC 2015: John Spray | The Ceph Storage SystemOSDC 2015: John Spray | The Ceph Storage System
OSDC 2015: John Spray | The Ceph Storage System
 
Scale 10x 01:22:12
Scale 10x 01:22:12Scale 10x 01:22:12
Scale 10x 01:22:12
 
XenSummit - 08/28/2012
XenSummit - 08/28/2012XenSummit - 08/28/2012
XenSummit - 08/28/2012
 
Strata - 03/31/2012
Strata - 03/31/2012Strata - 03/31/2012
Strata - 03/31/2012
 
Crimson: Ceph for the Age of NVMe and Persistent Memory
Crimson: Ceph for the Age of NVMe and Persistent MemoryCrimson: Ceph for the Age of NVMe and Persistent Memory
Crimson: Ceph for the Age of NVMe and Persistent Memory
 
Red Hat Gluster Storage Performance
Red Hat Gluster Storage PerformanceRed Hat Gluster Storage Performance
Red Hat Gluster Storage Performance
 
The Hive Think Tank: Ceph + RocksDB by Sage Weil, Red Hat.
The Hive Think Tank: Ceph + RocksDB by Sage Weil, Red Hat.The Hive Think Tank: Ceph + RocksDB by Sage Weil, Red Hat.
The Hive Think Tank: Ceph + RocksDB by Sage Weil, Red Hat.
 
RBD: What will the future bring? - Jason Dillaman
RBD: What will the future bring? - Jason DillamanRBD: What will the future bring? - Jason Dillaman
RBD: What will the future bring? - Jason Dillaman
 
Open Source Storage at Scale: Ceph @ GRNET
Open Source Storage at Scale: Ceph @ GRNETOpen Source Storage at Scale: Ceph @ GRNET
Open Source Storage at Scale: Ceph @ GRNET
 

Recently uploaded

英国UN学位证,北安普顿大学毕业证书1:1制作
英国UN学位证,北安普顿大学毕业证书1:1制作英国UN学位证,北安普顿大学毕业证书1:1制作
英国UN学位证,北安普顿大学毕业证书1:1制作qr0udbr0
 
How to submit a standout Adobe Champion Application
How to submit a standout Adobe Champion ApplicationHow to submit a standout Adobe Champion Application
How to submit a standout Adobe Champion ApplicationBradBedford3
 
Unveiling Design Patterns: A Visual Guide with UML Diagrams
Unveiling Design Patterns: A Visual Guide with UML DiagramsUnveiling Design Patterns: A Visual Guide with UML Diagrams
Unveiling Design Patterns: A Visual Guide with UML DiagramsAhmed Mohamed
 
Introduction Computer Science - Software Design.pdf
Introduction Computer Science - Software Design.pdfIntroduction Computer Science - Software Design.pdf
Introduction Computer Science - Software Design.pdfFerryKemperman
 
Ahmed Motair CV April 2024 (Senior SW Developer)
Ahmed Motair CV April 2024 (Senior SW Developer)Ahmed Motair CV April 2024 (Senior SW Developer)
Ahmed Motair CV April 2024 (Senior SW Developer)Ahmed Mater
 
Powering Real-Time Decisions with Continuous Data Streams
Powering Real-Time Decisions with Continuous Data StreamsPowering Real-Time Decisions with Continuous Data Streams
Powering Real-Time Decisions with Continuous Data StreamsSafe Software
 
Folding Cheat Sheet #4 - fourth in a series
Folding Cheat Sheet #4 - fourth in a seriesFolding Cheat Sheet #4 - fourth in a series
Folding Cheat Sheet #4 - fourth in a seriesPhilip Schwarz
 
CRM Contender Series: HubSpot vs. Salesforce
CRM Contender Series: HubSpot vs. SalesforceCRM Contender Series: HubSpot vs. Salesforce
CRM Contender Series: HubSpot vs. SalesforceBrainSell Technologies
 
Open Source Summit NA 2024: Open Source Cloud Costs - OpenCost's Impact on En...
Open Source Summit NA 2024: Open Source Cloud Costs - OpenCost's Impact on En...Open Source Summit NA 2024: Open Source Cloud Costs - OpenCost's Impact on En...
Open Source Summit NA 2024: Open Source Cloud Costs - OpenCost's Impact on En...Matt Ray
 
Intelligent Home Wi-Fi Solutions | ThinkPalm
Intelligent Home Wi-Fi Solutions | ThinkPalmIntelligent Home Wi-Fi Solutions | ThinkPalm
Intelligent Home Wi-Fi Solutions | ThinkPalmSujith Sukumaran
 
Comparing Linux OS Image Update Models - EOSS 2024.pdf
Comparing Linux OS Image Update Models - EOSS 2024.pdfComparing Linux OS Image Update Models - EOSS 2024.pdf
Comparing Linux OS Image Update Models - EOSS 2024.pdfDrew Moseley
 
Call Us🔝>༒+91-9711147426⇛Call In girls karol bagh (Delhi)
Call Us🔝>༒+91-9711147426⇛Call In girls karol bagh (Delhi)Call Us🔝>༒+91-9711147426⇛Call In girls karol bagh (Delhi)
Call Us🔝>༒+91-9711147426⇛Call In girls karol bagh (Delhi)jennyeacort
 
Sending Calendar Invites on SES and Calendarsnack.pdf
Sending Calendar Invites on SES and Calendarsnack.pdfSending Calendar Invites on SES and Calendarsnack.pdf
Sending Calendar Invites on SES and Calendarsnack.pdf31events.com
 
Post Quantum Cryptography – The Impact on Identity
Post Quantum Cryptography – The Impact on IdentityPost Quantum Cryptography – The Impact on Identity
Post Quantum Cryptography – The Impact on Identityteam-WIBU
 
Odoo 14 - eLearning Module In Odoo 14 Enterprise
Odoo 14 - eLearning Module In Odoo 14 EnterpriseOdoo 14 - eLearning Module In Odoo 14 Enterprise
Odoo 14 - eLearning Module In Odoo 14 Enterprisepreethippts
 
SpotFlow: Tracking Method Calls and States at Runtime
SpotFlow: Tracking Method Calls and States at RuntimeSpotFlow: Tracking Method Calls and States at Runtime
SpotFlow: Tracking Method Calls and States at Runtimeandrehoraa
 
Salesforce Implementation Services PPT By ABSYZ
Salesforce Implementation Services PPT By ABSYZSalesforce Implementation Services PPT By ABSYZ
Salesforce Implementation Services PPT By ABSYZABSYZ Inc
 
Software Project Health Check: Best Practices and Techniques for Your Product...
Software Project Health Check: Best Practices and Techniques for Your Product...Software Project Health Check: Best Practices and Techniques for Your Product...
Software Project Health Check: Best Practices and Techniques for Your Product...Velvetech LLC
 
Innovate and Collaborate- Harnessing the Power of Open Source Software.pdf
Innovate and Collaborate- Harnessing the Power of Open Source Software.pdfInnovate and Collaborate- Harnessing the Power of Open Source Software.pdf
Innovate and Collaborate- Harnessing the Power of Open Source Software.pdfYashikaSharma391629
 
Unveiling the Future: Sylius 2.0 New Features
Unveiling the Future: Sylius 2.0 New FeaturesUnveiling the Future: Sylius 2.0 New Features
Unveiling the Future: Sylius 2.0 New FeaturesŁukasz Chruściel
 

Recently uploaded (20)

英国UN学位证,北安普顿大学毕业证书1:1制作
英国UN学位证,北安普顿大学毕业证书1:1制作英国UN学位证,北安普顿大学毕业证书1:1制作
英国UN学位证,北安普顿大学毕业证书1:1制作
 
How to submit a standout Adobe Champion Application
How to submit a standout Adobe Champion ApplicationHow to submit a standout Adobe Champion Application
How to submit a standout Adobe Champion Application
 
Unveiling Design Patterns: A Visual Guide with UML Diagrams
Unveiling Design Patterns: A Visual Guide with UML DiagramsUnveiling Design Patterns: A Visual Guide with UML Diagrams
Unveiling Design Patterns: A Visual Guide with UML Diagrams
 
Introduction Computer Science - Software Design.pdf
Introduction Computer Science - Software Design.pdfIntroduction Computer Science - Software Design.pdf
Introduction Computer Science - Software Design.pdf
 
Ahmed Motair CV April 2024 (Senior SW Developer)
Ahmed Motair CV April 2024 (Senior SW Developer)Ahmed Motair CV April 2024 (Senior SW Developer)
Ahmed Motair CV April 2024 (Senior SW Developer)
 
Powering Real-Time Decisions with Continuous Data Streams
Powering Real-Time Decisions with Continuous Data StreamsPowering Real-Time Decisions with Continuous Data Streams
Powering Real-Time Decisions with Continuous Data Streams
 
Folding Cheat Sheet #4 - fourth in a series
Folding Cheat Sheet #4 - fourth in a seriesFolding Cheat Sheet #4 - fourth in a series
Folding Cheat Sheet #4 - fourth in a series
 
CRM Contender Series: HubSpot vs. Salesforce
CRM Contender Series: HubSpot vs. SalesforceCRM Contender Series: HubSpot vs. Salesforce
CRM Contender Series: HubSpot vs. Salesforce
 
Open Source Summit NA 2024: Open Source Cloud Costs - OpenCost's Impact on En...
Open Source Summit NA 2024: Open Source Cloud Costs - OpenCost's Impact on En...Open Source Summit NA 2024: Open Source Cloud Costs - OpenCost's Impact on En...
Open Source Summit NA 2024: Open Source Cloud Costs - OpenCost's Impact on En...
 
Intelligent Home Wi-Fi Solutions | ThinkPalm
Intelligent Home Wi-Fi Solutions | ThinkPalmIntelligent Home Wi-Fi Solutions | ThinkPalm
Intelligent Home Wi-Fi Solutions | ThinkPalm
 
Comparing Linux OS Image Update Models - EOSS 2024.pdf
Comparing Linux OS Image Update Models - EOSS 2024.pdfComparing Linux OS Image Update Models - EOSS 2024.pdf
Comparing Linux OS Image Update Models - EOSS 2024.pdf
 
Call Us🔝>༒+91-9711147426⇛Call In girls karol bagh (Delhi)
Call Us🔝>༒+91-9711147426⇛Call In girls karol bagh (Delhi)Call Us🔝>༒+91-9711147426⇛Call In girls karol bagh (Delhi)
Call Us🔝>༒+91-9711147426⇛Call In girls karol bagh (Delhi)
 
Sending Calendar Invites on SES and Calendarsnack.pdf
Sending Calendar Invites on SES and Calendarsnack.pdfSending Calendar Invites on SES and Calendarsnack.pdf
Sending Calendar Invites on SES and Calendarsnack.pdf
 
Post Quantum Cryptography – The Impact on Identity
Post Quantum Cryptography – The Impact on IdentityPost Quantum Cryptography – The Impact on Identity
Post Quantum Cryptography – The Impact on Identity
 
Odoo 14 - eLearning Module In Odoo 14 Enterprise
Odoo 14 - eLearning Module In Odoo 14 EnterpriseOdoo 14 - eLearning Module In Odoo 14 Enterprise
Odoo 14 - eLearning Module In Odoo 14 Enterprise
 
SpotFlow: Tracking Method Calls and States at Runtime
SpotFlow: Tracking Method Calls and States at RuntimeSpotFlow: Tracking Method Calls and States at Runtime
SpotFlow: Tracking Method Calls and States at Runtime
 
Salesforce Implementation Services PPT By ABSYZ
Salesforce Implementation Services PPT By ABSYZSalesforce Implementation Services PPT By ABSYZ
Salesforce Implementation Services PPT By ABSYZ
 
Software Project Health Check: Best Practices and Techniques for Your Product...
Software Project Health Check: Best Practices and Techniques for Your Product...Software Project Health Check: Best Practices and Techniques for Your Product...
Software Project Health Check: Best Practices and Techniques for Your Product...
 
Innovate and Collaborate- Harnessing the Power of Open Source Software.pdf
Innovate and Collaborate- Harnessing the Power of Open Source Software.pdfInnovate and Collaborate- Harnessing the Power of Open Source Software.pdf
Innovate and Collaborate- Harnessing the Power of Open Source Software.pdf
 
Unveiling the Future: Sylius 2.0 New Features
Unveiling the Future: Sylius 2.0 New FeaturesUnveiling the Future: Sylius 2.0 New Features
Unveiling the Future: Sylius 2.0 New Features
 

Community Update at OpenStack Summit Boston

  • 1. COMMUNITY UPDATE SAGE WEIL – RED HAT 2017.05.11
  • 2. 2 UPSTREAM RELEASES Jewel (LTS) Spring 2016 Kraken Fall 2016 Luminous (LTS) Spring 2017 12.2.z 10.2.z 14.2.z WE ARE HERE Mimic Fall 2017? Nautilus? (LTS) Spring 2018?
  • 4. 4 BLUESTORE: STABLE AND DEFAULT ● New OSD backend – consumes raw block device(s) – no more XFS – embeds rocksdb for metadata ● Fast on both HDDs (~2x) and SSDs (~1.5x) – Similar to FileStore on NVMe, where the device is not the bottleneck ● Smaller journals – happily uses fast SSD partition(s) for internal metadata, or NVRAM for journal ● Full data checksums (crc32c, xxhash, etc.) ● Inline compression (zlib, snappy) – policy driven by global or per-pool config, and/or client hints ● Stable and default
  • 5. 5 HDD: RANDOM WRITE 4 8 16 32 64 128 256 512 1024 2048 4096 0 100 200 300 400 500 Bluestore vs Filestore HDD Random Write Throughput Filestore Bluestore (wip-bitmap-alloc-perf) BS (Master-a07452d) BS (Master-d62a4948) Bluestore (wip-bluestore-dw) IO Size Throughput(MB/s) 4 8 16 32 64 128 256 512 1024 2048 4096 0 500 1000 1500 2000 Bluestore vs Filestore HDD Random Write IOPS Filestore Bluestore (wip-bitmap-alloc-perf) BS (Master-a07452d) BS (Master-d62a4948) Bluestore (wip-bluestore-dw) IO Size IOPS
  • 6. 6 HDD: MIXED READ/WRITE 4 8 16 32 64 128 256 512 1024 2048 4096 0 50 100 150 200 250 300 350 Bluestore vs Filestore HDD Random RWThroughput Filestore Bluestore (wip-bitmap-alloc-perf) BS (Master-a07452d) BS (Master-d62a4948) Bluestore (wip-bluestore-dw) IO Size Throughput(MB/s) 4 8 16 32 64 128 256 512 1024 2048 4096 0 200 400 600 800 1000 1200 Bluestore vs Filestore HDD Random RW IOPS Filestore Bluestore (wip-bitmap-alloc-perf) BS (Master-a07452d) BS (Master-d62a4948) Bluestore (wip-bluestore-dw) IO Size IOPS
  • 7. 7 RGW ON HDD+NVME, EC 4+2 1 Bucket 128 Buckets 4 Buckets 512 Buckets Rados 1 RGW Server 4 RGW Servers Bench 0 200 400 600 800 1000 1200 1400 1600 1800 4+2 Erasure Coding RadosGW Write Tests 32MB Objects, 24 HDD/NVMe OSDs on 4 Servers, 4 Clients Filestore 512KB Chunks Filestore 4MB Chunks Bluestore 512KB Chunks Bluestore 4MB Chunks Throughput(MB/s)
  • 8. 8 RBD OVER ERASURE CODED POOLS ● aka erasure code overwrites ● requires BlueStore to perform reasonably ● significant improvement in efficiency over 3x replication – 2+2 → 2x 4+2 → 1.5x ● small writes slower than replication – early testing showed 4+2 is about half as fast as 3x replication ● large writes faster than replication – less IO to device ● implementation still does the “simple” thing – all writes update a full stripe
  • 9. 9 CEPH-MGR ● ceph-mgr – new management daemon to supplement ceph-mon (monitor) – easier integration point for python management logic – integrated metrics ● make ceph-mon scalable again – offload pg stats from mon to mgr – push to 10K OSDs (planned “big bang 3” @ CERN) ● new REST API – pecan – based on previous Calamari API ● built-in web dashboard – webby equivalent of 'ceph -s' M G ??? (time for new iconography)
  • 10. 10
  • 11. 11 ASYNCMESSENGER ● new network Messenger implementation – event driven – fixed-size thread pool ● RDMA backend (ibverbs) – built by default – limited testing, but seems stable! ● DPDK backend – prototype!
  • 12. 12 PERFECTLY BALANCED OSDS (FINALLY) ● CRUSH choose_args – alternate weight sets for individual rules – complete flexibility to optimize weights etc – fixes two problems ● imbalance – run numeric optimization to adjust weights to balance PG distribution for a pool (or cluster) ● multipick anomaly – adjust weights per position to correct for low- weighted devices (e.g., mostly empty rack) – backward compatible with pre-luminous clients for imbalance case ● pg upmap – explicitly map individual PGs to specific devices in OSDMap – simple offline optimizer balance PGs – by pool or by cluster – requires luminous+ clients
  • 13. 13 RADOS MISC ● CRUSH device classes – mark OSDs with class (hdd, ssd, etc) – out-of-box rules to map to specific class of devices within the same hierarchy ● streamlined disk replacement ● require_min_compat_client – simpler, safer configuration ● annotated/documented config options ● client backoff on stuck PGs or objects ● better EIO handling ● peering and recovery speedups ● fast OSD failure detection
  • 14. 14 S3 Swift Erasure coding Multisite federation Multisite replication NFS Encryption Tiering Deduplication RADOSGW Compression
  • 15. 15 ZONE CZONE B RGW METADATA SEARCH RADOSGW LIBRADOS M CLUSTER A MM M CLUSTER B MM RADOSGW RADOSGW RADOSGW LIBRADOS LIBRADOS LIBRADOS REST RADOSGW LIBRADOS M CLUSTER C MM REST ZONE A
  • 16. 16 RGW MISC ● NFS gateway – NFSv4 and v3 – full object access (not general purpose!) ● dynamic bucket index sharding – automatic (finally!) ● inline compression ● encryption – follows S3 encryption APIs ● S3 and Swift API odds and ends RADOSGW LIBRADOS
  • 17. 17 Erasure coding Multisite mirroring Persistent client cache Consistency groups Encryption iSCSI Trash RBD
  • 18. 18 RBD ● RBD over erasure coded pool – rbd create --data-pool <ecpoolname> ... ● RBD mirroring improvements – cooperative HA daemons – improved Cinder integration ● iSCSI – LIO tcmu-runner, librbd (full feature set) ● Kernel RBD improvements – exclusive locking, object map
  • 19. RADOS A reliable, autonomous, distributed object store comprised of self-healing, self-managing, intelligent storage nodes RADOS A reliable, autonomous, distributed object store comprised of self-healing, self-managing, intelligent storage nodes LIBRADOS A library allowing apps to directly access RADOS, with support for C, C++, Java, Python, Ruby, and PHP LIBRADOS A library allowing apps to directly access RADOS, with support for C, C++, Java, Python, Ruby, and PHP RBD A reliable and fully- distributed block device, with a Linux kernel client and a QEMU/KVM driver RBD A reliable and fully- distributed block device, with a Linux kernel client and a QEMU/KVM driver RADOSGW A bucket-based REST gateway, compatible with S3 and Swift RADOSGW A bucket-based REST gateway, compatible with S3 and Swift APPAPP APPAPP HOST/VMHOST/VM CLIENTCLIENT CEPH FS A POSIX-compliant distributed file system, with a Linux kernel client and support for FUSE CEPH FS A POSIX-compliant distributed file system, with a Linux kernel client and support for FUSE NEARLY AWESOME AWESOMEAWESOME AWESOME AWESOME
  • 20. 20 2017 = RGW A web services gateway for object storage, compatible with S3 and Swift LIBRADOS A library allowing apps to directly access RADOS (C, C++, Java, Python, Ruby, PHP) RADOS A software-based, reliable, autonomous, distributed object store comprised of self-healing, self-managing, intelligent storage nodes and lightweight monitors RBD A reliable, fully-distributed block device with cloud platform integration CEPHFS A distributed file system with POSIX semantics and scale-out metadata management OBJECT BLOCK FILE FULLY AWESOME
  • 21. 21 CEPHFS ● multiple active MDS daemons (finally!) ● subtree pinning to specific daemon ● directory fragmentation on by default – (snapshots still off by default) ● so many tests ● so many bugs fixed ● kernel client improvements
  • 24. 24 RADOS ● Peering optimizations ● IO path refactor and optimization – async, state-driven, futures – painful but necessary ● BlueStore and rocksdb optimization – rocksdb level0 compaction – alternative KV store? ● Erasure coding plugin API improvements – new codes with less IO for single-OSD failures DEVICE OSD OBJECTSTORE MESSENGER
  • 25. 25 QUALITY OF SERVICE ● Ongoing background development – dmclock distributed QoS queuing – minimum reservations and priority weighting ● Range of policies – IO type (background, client) – pool-based – client-based ● Theory is complex ● Prototype is promising, despite simplicity ● Missing management framework
  • 26. 26
  • 27. 27 CEPH STORAGE CLUSTER TIERING ● new RADOS ‘redirect’ primitive – basically a symlink, transparent to librados – replace “sparse” cache tier with base pool “index” APPLICATION BASE POOL (REPLICATED, SSD) APPLICATION CACHE POOL (REPLICATED) BASE POOL (HDD AND/OR ERASURE) CEPH STORAGE CLUSTER SLOW 1 (EC) SLOW #1 (...)
  • 28. 28 CEPH STORAGE CLUSTER DEDUPLICATION WIP ● Generalize redirect to a “manifest” – map of offsets to object “fragments” (vs a full object copy) ● Break objects into chunks – fixed size, or content fingerprint ● Store chunks in content-addressable pool – name object by sha256(content) – reference count chunks ● TBD – inline or post? – policy (when to promote inline, etc.) – agent home (embed in OSD, …) APPLICATION BASE POOL (REPLICATED, SSD) SLOW 1 (EC) CAS DEDUP POOL
  • 29. 29 CEPH-MGR: METRICS, MGMT ● Metrics aggregation – short-term time series out of the box – no persistent state – streaming to external platform (Prometheus, …) ● Host self-management functions – automatic CRUSH optimization – identification of slow devices – steer IO away from busy devices – device failure prediction M G consistency metrics summary
  • 30. 30 ARM ● aarch64 builds – centos7, ubuntu xenial – have some, but awaiting more build hardware in community lab ● thank you to partners! ● ppc64 ● armv7l – http://ceph.com/community/500-osd-ceph-cluster/
  • 31. 31 Gen 2 Gen3 ARM 32bit ARM 64 bit uBoot UEFI 1GB DDR Up to 4GB DDR 2x 1GbE SGMII 2x 2.5GbE SGMII Basic Security Enhanced Security No Flash option Flash option I2C Management Integrated Management Planning Gen3 availability to PoC partners Contact james.wilshire@wdc.com WDLABS MICROSERVER UPDATE
  • 32. 32 CLIENT CACHES! ● RGW – persistent read-only cache on NVMe – fully consistent (only caches immutable “tail” rados objects) – Mass Open Cloud ● RBD – persistent read-only cache of immutable clone parent images – writeback cache for improving write latency ● cluster image remains crash-consistent if client cache is lost ● CephFS – kernel client already uses kernel fscache facility
  • 34. 34 ● Red Hat ● Mirantis ● SUSE ● ZTE ● China Mobile ● XSky ● Digiware ● Intel ● Kylin Cloud ● Easystack ● Istuary Innovation Group ● Quantum ● Mellanox ● H3C ● Quantum ● UnitedStack ● Deutsche Telekom ● Reliance Jio Infocomm ● OVH ● Alibaba ● DreamHost ● CERN GROWING DEVELOPER COMMUNITY
  • 35. 35 ● Red Hat ● Mirantis ● SUSE ● ZTE ● China Mobile ● XSky ● Digiware ● Intel ● Kylin Cloud ● Easystack ● Istuary Innovation Group ● Quantum ● Mellanox ● H3C ● Quantum ● UnitedStack ● Deutsche Telekom ● Reliance Jio Infocomm ● OVH ● Alibaba ● DreamHost ● CERN OPENSTACK VENDORS
  • 36. 36 ● Red Hat ● Mirantis ● SUSE ● ZTE ● China Mobile ● XSky ● Digiware ● Intel ● Kylin Cloud ● Easystack ● Istuary Innovation Group ● Quantum ● Mellanox ● H3C ● Quantum ● UnitedStack ● Deutsche Telekom ● Reliance Jio Infocomm ● OVH ● Alibaba ● DreamHost ● CERN CLOUD OPERATORS
  • 37. 37 ● Red Hat ● Mirantis ● SUSE ● ZTE ● China Mobile ● XSky ● Digiware ● Intel ● Kylin Cloud ● Easystack ● Istuary Innovation Group ● Quantum ● Mellanox ● H3C ● Quantum ● UnitedStack ● Deutsche Telekom ● Reliance Jio Infocomm ● OVH ● Alibaba ● DreamHost ● CERN HARDWARE AND SOLUTION VENDORS
  • 38. 38 ● Red Hat ● Mirantis ● SUSE ● ZTE ● China Mobile ● XSky ● Digiware ● Intel ● Kylin Cloud ● Easystack ● Istuary Innovation Group ● Quantum ● Mellanox ● H3C ● Quantum ● UnitedStack ● Deutsche Telekom ● Reliance Jio Infocomm ● OVH ● Alibaba ● DreamHost ● CERN ???
  • 39. 39 ● Mailing list and IRC – http://ceph.com/IRC ● Ceph Developer Monthly – first Weds of every month – video conference (Bluejeans) – alternating APAC- and EMEA- friendly times ● Github – https://github.com/ceph/ ● Ceph Days – http://ceph.com/cephdays/ ● Meetups – http://ceph.com/meetups ● Ceph Tech Talks – http://ceph.com/ceph-tech-talks / ● ‘Ceph’ Youtube channel – (google it) ● Twitter – @ceph GET INVOLVED
  • 40. THANK YOU! Sage Weil CEPH PRINCIPAL ARCHITECT sage@redhat.com @liewegas