暂无分享,去创建一个
[1] Joan Bruna,et al. Exploiting Linear Structure Within Convolutional Networks for Efficient Evaluation , 2014, NIPS.
[2] Yoshua Bengio,et al. Training deep neural networks with low precision multiplications , 2014 .
[3] Ivan V. Oseledets,et al. Rectangular maximum-volume submatrices and their applications , 2015, ArXiv.
[4] Pritish Narayanan,et al. Deep Learning with Limited Numerical Precision , 2015, ICML.
[5] Bo Chen,et al. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.
[6] Jian Sun,et al. Accelerating Very Deep Convolutional Networks for Classification and Detection , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[7] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[8] Ivan V. Oseledets,et al. Speeding-up Convolutional Neural Networks Using Fine-tuned CP-Decomposition , 2014, ICLR.
[9] Danny C. Sorensen,et al. Nonlinear Model Reduction via Discrete Empirical Interpolation , 2010, SIAM J. Sci. Comput..
[10] Shinichi Nakajima,et al. Global analytic solution of fully-observed variational Bayesian matrix factorization , 2013, J. Mach. Learn. Res..
[11] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[12] Tao Zhang,et al. Model Compression and Acceleration for Deep Neural Networks: The Principles, Progress, and Challenges , 2018, IEEE Signal Processing Magazine.
[13] Nikos Komodakis,et al. Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer , 2016, ICLR.
[14] Jianxin Wu,et al. ThiNet: Pruning CNN Filters for a Thinner Net , 2019, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[15] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[16] David Duvenaud,et al. Neural Ordinary Differential Equations , 2018, NeurIPS.
[17] F. Michael,et al. PerforatedCNNs: Acceleration through Elimination of Redundant Convolutions , 2016, ICLR 2016.
[18] Yoshua Bengio,et al. FitNets: Hints for Thin Deep Nets , 2014, ICLR.
[19] Jing Liu,et al. Discrimination-aware Channel Pruning for Deep Neural Networks , 2018, NeurIPS.
[20] Victor S. Lempitsky,et al. Fast ConvNets Using Group-Wise Brain Damage , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[21] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[22] Andrew Zisserman,et al. Speeding up Convolutional Neural Networks with Low Rank Expansions , 2014, BMVC.
[23] Andrzej Cichocki,et al. MUSCO: Multi-Stage Compression of neural networks , 2019 .
[24] Zhiqiang Shen,et al. Learning Efficient Convolutional Networks through Network Slimming , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[25] David Duvenaud,et al. FFJORD: Free-form Continuous Dynamics for Scalable Reversible Generative Models , 2018, ICLR.
[26] David P. Woodruff,et al. Frequent Directions: Simple and Deterministic Matrix Sketching , 2015, SIAM J. Comput..
[27] David P. Woodruff. Sketching as a Tool for Numerical Linear Algebra , 2014, Found. Trends Theor. Comput. Sci..
[28] Xiangyu Zhang,et al. Channel Pruning for Accelerating Very Deep Neural Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[29] Gleb Gusev,et al. Efficient Rectangular Maximal-Volume Algorithm for Rating Elicitation in Collaborative Filtering , 2016, 2016 IEEE 16th International Conference on Data Mining (ICDM).
[30] Gianluigi Rozza,et al. Reduced Order Methods for Modeling and Computational Reduction , 2013 .
[31] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.