Compressing Deep Neural Networks with Probabilistic Data Structures
暂无分享,去创建一个
[1] Burton H. Bloom,et al. Space/time trade-offs in hash coding with allowable errors , 1970, CACM.
[2] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[3] Babak Hassibi,et al. Second Order Derivatives for Network Pruning: Optimal Brain Surgeon , 1992, NIPS.
[4] Bernard Chazelle,et al. The Bloomier filter: an efficient data structure for static support lookup tables , 2004, SODA '04.
[5] David J. C. MacKay,et al. Information Theory, Inference, and Learning Algorithms , 2004, IEEE Transactions on Information Theory.
[6] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[7] Gu-Yeon Wei,et al. Minerva: Enabling Low-Power, Highly-Accurate Deep Neural Network Accelerators , 2016, 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA).
[8] Yurong Chen,et al. Dynamic Network Surgery for Efficient DNNs , 2016, NIPS.
[9] Alexander M. Rush,et al. Weightless: Lossy Weight Encoding For Deep Neural Network Compression , 2018, ICML.