Internet-scale storage systems under churn - A steady state analysis

Content storage in a distributed collaborative environment uses redundancy for better resilience. This redundancy is either achieved by pure replication or using erasure codes for more efficient utilization of available storage. Increasingly such schemes are employed in peer-to-peer environments, though more realistically in local or wide area network settings, which is still characterized by moderate attrition (churn) of network membership. An mportant aspect of guaranteeing persistence and availability in such a system is to maintain the redundancy. A proactive mechanism of replenishing lost redundancy can however be prohibitive, and hence lazier maintenance schemes are desirable. Such systems' performance has typically been studied based on simulations, and more recently some prototype based experimental results have also been reported. While such a methodology is essential in designing and deploying a reliable system, the missing link in the existing literature is a sufficiently accurate analytical model to capture the dynamics of the system. The existing analytical models essentially capture the static resilience of the system. What's more realistic is to evaluate whether - given a rate of churn, and some chosen maintenance strategy, the system operates at a steady state and if so, what is the operational cost of such a system. Since simulations are restricted to chosen workloads (be it synthetic or trace driven), they are limited to only those chosen workloads, even while providing a good initial (hopefully empirical) insight to systems designers. A precise analytical model capturing the system's dynamicity however is expected to provide a more objective insight, and is the primary contribution of this work.

[1]  Rajkumar Buyya,et al.  High Performance Mass Storage and Parallel I/O: Technologies and Applications , 2001 .

[2]  Ben Y. Zhao,et al.  OceanStore: an architecture for global-scale persistent storage , 2000, SIGP.

[3]  John Kubiatowicz,et al.  Erasure Coding Vs. Replication: A Quantitative Comparison , 2002, IPTPS.

[4]  Rodrigo Rodrigues,et al.  High Availability in DHTs: Erasure Coding vs. Replication , 2005, IPTPS.

[5]  Carl Staelin,et al.  The HP AutoRAID hierarchical storage system , 1995, SOSP.

[6]  Stefan Savage,et al.  Understanding Availability , 2003, IPTPS.

[7]  David R. Karger,et al.  Wide-area cooperative storage with CFS , 2001, SOSP.

[8]  Randy H. Katz,et al.  A case for redundant arrays of inexpensive disks (RAID) , 1988, SIGMOD '88.

[9]  Stefan Savage,et al.  Total Recall: System Support for Automated Availability Management , 2004, NSDI.

[10]  Francis Heylighen,et al.  Cybernetics and Second-Order Cybernetics , 2001 .

[11]  Rodrigo Rodrigues,et al.  Proceedings of Hotos Ix: the 9th Workshop on Hot Topics in Operating Systems Hotos Ix: the 9th Workshop on Hot Topics in Operating Systems High Availability, Scalable Storage, Dynamic Peer Networks: Pick Two , 2022 .

[12]  Ira Pramanick,et al.  High Availability , 2001, Int. J. High Perform. Comput. Appl..

[13]  David R. Karger,et al.  Analysis of the evolution of peer-to-peer systems , 2002, PODC '02.

[14]  Krishna P. Gummadi,et al.  The impact of DHT routing geometry on resilience and proximity , 2003, SIGCOMM '03.

[15]  Mary Baker,et al.  The LOCKSS peer-to-peer digital preservation system , 2005, TOCS.

[16]  Brian D. Noble,et al.  Proceedings of the 5th Symposium on Operating Systems Design and Implementation Pastiche: Making Backup Cheap and Easy , 2022 .

[17]  Ben Y. Zhao,et al.  Pond: The OceanStore Prototype , 2003, FAST.

[18]  Seif Haridi,et al.  A Physics-Inspired Performance Evaluation of a Structured Peer-to-Peer Overlay Network , 2005, Parallel and Distributed Computing and Networks.

[19]  Karl Aberer,et al.  Efficient, self-contained handling of identity in peer-to-peer systems , 2004, IEEE Transactions on Knowledge and Data Engineering.

[20]  David E. Culler,et al.  A blueprint for introducing disruptive technology into the Internet , 2003, CCRV.

[21]  Andreas Haeberlen,et al.  Glacier: highly durable, decentralized storage despite massive correlated failures , 2005, NSDI.

[22]  Michael Luby,et al.  LT codes , 2002, The 43rd Annual IEEE Symposium on Foundations of Computer Science, 2002. Proceedings..