Behaviors of Storage Backends in Ceph Object Store

Ceph is a scalable, reliable and high-performance storage solution that is widely used in the cloud computing environment. Internally, Ceph provides three different storage backends: FileStore, KStore and BlueStore. However, little effort has been devoted to identifying the differences in those storage backends and their implications on performance. In this paper, we carry out extensive analysis with a microbenchmark and a long-term workload to compare Ceph storage backends and understand their write behaviors by focusing on WAF (Write Amplification Factor). To accurately analyze WAF, we carefully classify write traffic into several categories for each storage backend. We find that writes are amplified by more than 13x, no matter which Ceph storage backend is used. In FileStore, the overhead of Ceph write-ahead journaling triples write traffic compared to the original data size. Also, FileStore has the journaling of journal problem, generating a relatively large amount of file system metadata and journal traffic. KStore suffers severe fluctuations in IOPS (I/O Operations Per Second) and WAF due to large compaction overheads. BlueStore shows the stable performance on both HDDs and SSDs in terms of IOPS, WAF and latency. Overall, FileStore performs the best among all storage backends on SSDs, while BlueStore is also highly promising with good average and tail latency even on HDDs.

[1]  Chao-Tung Yang,et al.  Implementation of a Software-Defined Storage Service with Heterogeneous Storage Technologies , 2015, 2015 IEEE 29th International Conference on Advanced Information Networking and Applications Workshops.

[2]  A. T. Chronopoulos,et al.  Ceph Distributed File System Benchmarks on an Openstack Cloud , 2015, 2015 IEEE International Conference on Cloud Computing in Emerging Markets (CCEM).

[3]  Irfan Ahmad Easy and Efficient Disk I/O Workload Characterization in VMware ESX Server , 2007, 2007 IEEE 10th International Symposium on Workload Characterization.

[4]  Meng Zhu,et al.  Journaling of journal is (almost) free , 2014, FAST.

[5]  Heon Young Yeom,et al.  Performance Optimization for All Flash Scale-Out Storage , 2016, 2016 IEEE International Conference on Cluster Computing (CLUSTER).

[6]  Carlos Maltzahn,et al.  Ceph: a scalable, high-performance distributed file system , 2006, OSDI '06.

[7]  Achim Streit,et al.  Evaluating the performance and scalability of the Ceph distributed storage system , 2014, 2014 IEEE International Conference on Big Data (Big Data).

[8]  Feiyi Wang,et al.  Performance and scalability evaluation of the Ceph parallel file system , 2013, PDSW@SC.

[9]  S.A. Brandt,et al.  CRUSH: Controlled, Scalable, Decentralized Placement of Replicated Data , 2006, ACM/IEEE SC 2006 Conference (SC'06).