Managing the Data Replicas Efficiently in Time-Zoned Multilayer Chord

Replica is better than erasure coding when managing the redundancy of a storage system under high churn. It needs extra network communication to keep the target redundancy. By classifying the hosts into different categories of availability and placing the data replicas carefully may reduce the maintenance overhead. Considering a globe scale P2P system, the time-zoned multilayer Chord divides the hosts into 24 categories according to the time zone where they located. All the hosts are belongs to a Main Chord which consisted of all the hosts and one Local Chord which consisted of the hosts located in the same time zone. Because of the daily behavior of the people in same time zone are nearly similar, which means their computers are more likely to be available in the daytime and less likely to be available in the nighttime. So, for a certain time, there is nearly half of the hosts get higher availability than the other hosts. As the hosts in the same time zone are considered, their availability evaluated by the system is shifting periodically in a 24 hours cycle. By give a carefully placement of the data replicas on this two layered Chord, the less communication is needed to maintain the same target availability. The experiments show that dividing hosts by location of time zone might reduce the overhead of redundancy management.

[1]  Andreas Haeberlen,et al.  Efficient Replica Maintenance for Distributed Storage Systems , 2006, NSDI.

[2]  Daniel Stutzbach,et al.  Understanding churn in peer-to-peer networks , 2006, IMC '06.

[3]  Taieb Znati,et al.  Modeling Churn in P2P Networks , 2007, 40th Annual Simulation Symposium (ANSS'07).

[4]  Yun Li,et al.  A Churn Model Based on the Global Geographical Distribution of Nodes , 2010, 2010 Ninth International Symposium on Distributed Computing and Applications to Business, Engineering and Science.

[5]  Gade Krishna,et al.  A scalable peer-to-peer lookup protocol for Internet applications , 2012 .

[6]  Stefan Savage,et al.  Total Recall: System Support for Automated Availability Management , 2004, NSDI.

[7]  John Kubiatowicz,et al.  Handling churn in a DHT , 2004 .

[8]  D. M. Chiu,et al.  Erasure code replication revisited , 2004, Proceedings. Fourth International Conference on Peer-to-Peer Computing, 2004. Proceedings..

[9]  Aaas News,et al.  Book Reviews , 1893, Buffalo Medical and Surgical Journal.

[10]  Jacky C. Chu,et al.  Availability and locality measurements of peer-to-peer file systems , 2002, SPIE ITCom.

[11]  John Kubiatowicz,et al.  Erasure Coding Vs. Replication: A Quantitative Comparison , 2002, IPTPS.

[12]  David R. Karger,et al.  Wide-area cooperative storage with CFS , 2001, SOSP.