暂无分享,去创建一个
[1] Michael Carbin,et al. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks , 2018, ICLR.
[2] Bruno A. Olshausen,et al. The Sparse Manifold Transform , 2018, NeurIPS.
[3] Changshui Zhang,et al. Sparse DNNs with Improved Adversarial Robustness , 2018, NeurIPS.
[4] Yaser S. Abu-Mostafa,et al. On the K-Winners-Take-All Network , 1988, NIPS.
[5] Honglak Lee,et al. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations , 2009, ICML '09.
[6] David J. Field,et al. Sparse coding with an overcomplete basis set: A strategy employed by V1? , 1997, Vision Research.
[7] Lawrence D. Jackel,et al. Backpropagation Applied to Handwritten Zip Code Recognition , 1989, Neural Computation.
[8] Philip H. S. Torr,et al. SNIP: Single-shot Network Pruning based on Connection Sensitivity , 2018, ICLR.
[9] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[11] Honglak Lee,et al. Sparse deep belief net model for visual area V2 , 2007, NIPS.
[12] Subutai Ahmad,et al. How do neurons operate on sparse distributed representations? A mathematical theory of sparsity, neurons and active dendrites , 2016, ArXiv.
[13] J. Dunning. The elephant in the room. , 2013, European journal of cardio-thoracic surgery : official journal of the European Association for Cardio-thoracic Surgery.
[14] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[15] Jürgen Schmidhuber,et al. Compete to Compute , 2013, NIPS.
[16] Yuwei Cui,et al. The HTM Spatial Pooler—A Neocortical Algorithm for Online Sparse Distributed Coding , 2016, bioRxiv.
[17] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[18] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[19] Brendan J. Frey,et al. Winner-Take-All Autoencoders , 2014, NIPS.
[20] Hanan Samet,et al. Pruning Filters for Efficient ConvNets , 2016, ICLR.
[21] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[22] Jimmy J. Lin,et al. Deep Residual Learning for Small-Footprint Keyword Spotting , 2017, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[23] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[24] Pentti Kanerva,et al. Sparse Distributed Memory , 1988 .
[25] Brendan J. Frey,et al. k-Sparse Autoencoders , 2013, ICLR.
[26] Gideon Kowadlo,et al. Sparse Unsupervised Capsules Generalize Better , 2018, ArXiv.
[27] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[28] Geoffrey E. Hinton,et al. 3D Object Recognition with Deep Belief Nets , 2009, NIPS.
[29] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.
[30] Tara N. Sainath,et al. Convolutional neural networks for small-footprint keyword spotting , 2015, INTERSPEECH.