Distributed File Sharing: Network Coding Meets Compressed Sensing

In a peer-to-peer file distribution network, a large file is split into blocks residing in multiple storage locations. A peer node tries to retrieve the original file by downloading blocks from randomly chosen peers. We compare the performance of four storage strategies: uncoded, erasure coding, random linear coding, and random linear coding over coded blocks. We show that, in principle, random linear coding makes a better tradeoff between the storage requirement and decoding complexity. However, the sparsity of the file blocks is not fully exploited by random linear combinations of all original blocks. Motivated by the recent results from compressed sensing, we study the design tradeoff in random linear coding over coded blocks and propose an efficient decoding algorithm based on basis pursuit. We show that the minimum number of storage locations that a peer note has to connect to reconstruct the entire file with high probability can be significantly smaller than the total number of blocks that the file is broken into

[1]  Michael Mitzenmacher,et al.  A digital fountain approach to asynchronous reliable multicast , 2002, IEEE J. Sel. Areas Commun..

[2]  Harald Niederreiter,et al.  Probability and computing: randomized algorithms and probabilistic analysis , 2006, Math. Comput..

[3]  Muriel Medard,et al.  How good is random linear coding based distributed networked storage , 2005 .

[4]  Rudolf Ahlswede,et al.  Network information flow , 2000, IEEE Trans. Inf. Theory.

[5]  Joel A. Tropp,et al.  Greed is good: algorithmic results for sparse approximation , 2004, IEEE Transactions on Information Theory.

[6]  Christos Gkantsidis,et al.  Network coding for large scale content distribution , 2005, Proceedings IEEE 24th Annual Joint Conference of the IEEE Computer and Communications Societies..

[7]  David J. C. MacKay,et al.  Information Theory, Inference, and Learning Algorithms , 2004, IEEE Transactions on Information Theory.