暂无分享,去创建一个
Matthieu Cord | Kevin Bailly | Arnaud Dapogny | Edouard Yvinec | M. Cord | Arnaud Dapogny | Kévin Bailly | Edouard Yvinec
[1] Rémi Gribonval,et al. And the Bit Goes Down: Revisiting the Quantization of Neural Networks , 2019, ICLR.
[2] Mark Sandler,et al. MobileNetV2: Inverted Residuals and Linear Bottlenecks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[3] Hanan Samet,et al. Pruning Filters for Efficient ConvNets , 2016, ICLR.
[4] S. P. Lloyd,et al. Least squares quantization in PCM , 1982, IEEE Trans. Inf. Theory.
[5] Hao Cheng,et al. Pruning Filter in Filter , 2020, NeurIPS.
[6] Erich Elsen,et al. Rigging the Lottery: Making All Tickets Winners , 2020, ICML.
[7] Zhijian Liu,et al. HAQ: Hardware-Aware Automated Quantization With Mixed Precision , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[8] Martin Jaggi,et al. Dynamic Model Pruning with Feedback , 2020, ICLR.
[9] Mingjie Sun,et al. Rethinking the Value of Network Pruning , 2018, ICLR.
[10] Yi Yang,et al. Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks , 2018, IJCAI.
[11] Quoc V. Le,et al. Searching for MobileNetV3 , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[12] Michael Carbin,et al. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks , 2018, ICLR.
[13] Liujuan Cao,et al. Towards Optimal Structured CNN Pruning via Generative Adversarial Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[14] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[15] Yuheng Huang,et al. Neuron-level Structured Pruning using Polarization Regularizer , 2020, NeurIPS.
[16] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[17] Erich Elsen,et al. The State of Sparsity in Deep Neural Networks , 2019, ArXiv.
[18] R. Venkatesh Babu,et al. Data-free Parameter Pruning for Deep Neural Networks , 2015, BMVC.
[19] Zhiru Zhang,et al. Improving Neural Network Quantization without Retraining using Outlier Channel Splitting , 2019, ICML.
[20] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[21] Iasonas Kokkinos,et al. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[22] Jun Lu,et al. An Equivalence of Fully Connected Layer and Convolutional Layer , 2017, ArXiv.
[23] Alexander Finkelstein,et al. Same, Same But Different - Recovering Neural Network Quantization Error Through Weight Factorization , 2019, ICML.
[24] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[26] Rongrong Ji,et al. HRank: Filter Pruning Using High-Rank Feature Map , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Michael Carbin,et al. Comparing Rewinding and Fine-tuning in Neural Network Pruning , 2019, ICLR.
[28] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[29] Daniel L. K. Yamins,et al. Pruning neural networks without any data by iteratively conserving synaptic flow , 2020, NeurIPS.
[30] Song Han,et al. EIE: Efficient Inference Engine on Compressed Deep Neural Network , 2016, 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA).
[31] Dan Feldman,et al. Provable Filter Pruning for Efficient Neural Networks , 2019, ICLR.
[32] Naiyan Wang,et al. Data-Driven Sparse Structure Selection for Deep Neural Networks , 2017, ECCV.
[33] Tao Zhang,et al. A Survey of Model Compression and Acceleration for Deep Neural Networks , 2017, ArXiv.
[34] Dacheng Tao,et al. SCOP: Scientific Control for Reliable Neural Network Pruning , 2020, NeurIPS.
[35] Jacek M. Zurada,et al. Redundant feature pruning for accelerated inference in deep neural networks , 2019, Neural Networks.
[36] Derek Hoiem,et al. Dreaming to Distill: Data-Free Knowledge Transfer via DeepInversion , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[37] Markus Nagel,et al. Data-Free Quantization Through Weight Equalization and Bias Correction , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[38] Jinwoo Shin,et al. Lookahead: a Far-Sighted Alternative of Magnitude-based Pruning , 2020, ICLR.
[39] Philip H. S. Torr,et al. A Signal Propagation Perspective for Pruning Neural Networks at Initialization , 2019, ICLR.
[40] Jianxin Wu,et al. ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[41] Suhyun Kim,et al. Neuron Merging: Compensating for Pruned Neurons , 2020, NeurIPS.