DeepZip: Lossless Data Compression Using Recurrent Neural Networks
暂无分享,去创建一个
Idoia Ochoa | Shubham Chandak | Kedar Tatwawadi | Mohit Goyal | Kedar Tatwawadi | Idoia Ochoa | Mohit Goyal | Shubham Chandak
[1] George Marsaglia,et al. Toward a universal random number generator , 1987 .
[2] Sang Joon Kim,et al. A Mathematical Theory of Communication , 2006 .
[3] David A. Huffman,et al. A method for the construction of minimum-redundancy codes , 1952, Proceedings of the IRE.
[4] M. Sawicki,et al. Human Genome Project. , 1993, American journal of surgery.
[5] David Cox. Syntactically Informed Text Compression with Recurrent Neural Networks , 2016, ArXiv.
[6] Ian H. Witten,et al. Arithmetic coding for data compression , 1987, CACM.
[7] Y. Shtarkov,et al. The context-tree weighting method: basic properties , 1995, IEEE Trans. Inf. Theory.
[8] Jürgen Schmidhuber,et al. Sequential neural text compression , 1996, IEEE Trans. Neural Networks.
[9] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[10] Jianxin Wu. Hidden Markov model , 2018 .
[11] Matthew V. Mahoney,et al. Fast Text Compression with Neural Networks , 2000, FLAIRS Conference.
[12] Armando J. Pinho,et al. Efficient Compression of Genomic Sequences , 2016, 2016 Data Compression Conference (DCC).
[14] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[15] Ian H. Witten,et al. Data Compression Using Adaptive Coding and Partial String Matching , 1984, IEEE Trans. Commun..
[16] Abraham Lempel,et al. A universal algorithm for sequential data compression , 1977, IEEE Trans. Inf. Theory.