View divergence control of replicated data using update delay estimation

We propose a method to control the view divergence of replicated data when copies of sites in a replicated database are asynchronously updated. The view divergence of the replicated data is the difference in the lateness of the updates reflected in the data acquired by clients. Our method accesses multiple sites and provides a client with data that reflects all the updates received by the sites. We first define the probabilistic lateness of updates reflected in acquired data as read data freshness (RDF). The degrees of RDF of data acquired by clients is the range of the view divergence. Second, we propose a way to select sites in a replicated database by using the probability distribution of the update delays so that the data acquired by a client satisfies its required RDF. This way calculates the minimum number of sites in order to reduce the overhead of read transactions. Our method continues to adaptively and reliably provide data that meet the client's requirements in an environment where the delay of update propagation varies and applications' requirements change depending on situations. Finally, we evaluated the view divergence we can feasibly control using our method. The evaluation is done by means of simulations. The evaluation shows that our method can feasibly control the view divergence to about 1/4 that of a normal read transaction.

[1]  S. Ono,et al.  A statistical method for time synchronization of computer clocks with precisely frequency-synchronized oscillators , 1998, Proceedings. 18th International Conference on Distributed Computing Systems (Cat. No.98CB36183).

[2]  Judah Levine,et al.  An algorithm to synchronize the time of a computer to universal time , 1995, TNET.

[3]  J. Barnes Statistical Analysis for Engineers and Scientists: A Computer-Based Approach , 1994 .

[4]  ShashaDennis,et al.  The dangers of replication and a solution , 1996 .

[5]  Liuba Shrira,et al.  Lazy replication: exploiting the semantics of distributed services , 1990, ACM SIGOPS European Workshop.

[6]  Liuba Shrira,et al.  Providing high availability using lazy replication , 1992, TOCS.

[7]  Michael J. Fischer,et al.  Sacrificing serializability to attain high availability of data in an unreliable network , 1982, PODS.

[8]  Judah Levine Precision synchronization of computer network clocks , 1997 .

[9]  Kenneth P. Birman,et al.  Probabilistic Broadcast , 1996 .

[10]  Esther Pacitti,et al.  Improving data freshness in lazy master schemes , 1998, Proceedings. 18th International Conference on Distributed Computing Systems (Cat. No.98CB36183).

[11]  Michael K. Reiter,et al.  Probabilistic quorum systems , 1997, PODC '97.

[12]  Andreas Reuter,et al.  Transaction Processing: Concepts and Techniques , 1992 .

[13]  Dennis Shasha,et al.  The dangers of replication and a solution , 1996, SIGMOD '96.

[14]  Kenneth P. Birman,et al.  Building Secure and Reliable Network Applications , 1996 .

[15]  Bharat K. Bhargava,et al.  Replication Techniques in Distributed Systems , 1996, Advances in Database Systems.