暂无分享,去创建一个
[1] Boris Ginsburg,et al. Factorization tricks for LSTM networks , 2017, ICLR.
[2] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[3] Andrew Zisserman,et al. Speeding up Convolutional Neural Networks with Low Rank Expansions , 2014, BMVC.
[4] Tara N. Sainath,et al. Learning compact recurrent neural networks , 2016, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[5] Yuan Yu,et al. TensorFlow: A system for large-scale machine learning , 2016, OSDI.
[6] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[7] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[8] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[9] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[10] Stephen J. Wright,et al. Hogwild: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent , 2011, NIPS.
[11] John Tran,et al. cuDNN: Efficient Primitives for Deep Learning , 2014, ArXiv.
[12] Rio Yokota,et al. Accelerating Matrix Multiplication in Deep Learning by Using Low-Rank Approximation , 2017, 2017 International Conference on High Performance Computing & Simulation (HPCS).
[13] Ivan V. Oseledets,et al. Speeding-up Convolutional Neural Networks Using Fine-tuned CP-Decomposition , 2014, ICLR.
[14] Dong Yu,et al. 1-bit stochastic gradient descent and its application to data-parallel distributed training of speech DNNs , 2014, INTERSPEECH.
[15] Tianbao Yang,et al. Improved Dropout for Shallow and Deep Learning , 2016, NIPS.
[16] Edith Cohen,et al. Approximating matrix multiplication for pattern recognition tasks , 1997, SODA '97.
[17] Christian Osendorfer,et al. On Fast Dropout and its Applicability to Recurrent Networks , 2013, ICLR.
[18] David P. Woodruff,et al. Numerical linear algebra in the streaming model , 2009, STOC '09.
[19] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[20] Petros Drineas,et al. Fast Monte Carlo Algorithms for Matrices I: Approximating Matrix Multiplication , 2006, SIAM J. Comput..
[21] Ran El-Yaniv,et al. Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations , 2016, J. Mach. Learn. Res..
[22] Christophe Garcia,et al. Simplifying ConvNets for Fast Learning , 2012, ICANN.
[23] Xu Sun,et al. Training Simplification and Model Simplification for Deep Learning : A Minimal Effort Back Propagation Method , 2017, IEEE Transactions on Knowledge and Data Engineering.
[24] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[25] Nikko Strom,et al. Scalable distributed DNN training using commodity GPU cloud computing , 2015, INTERSPEECH.
[26] Hao Wu,et al. Mixed Precision Training , 2017, ICLR.
[27] Cong Xu,et al. TernGrad: Ternary Gradients to Reduce Communication in Distributed Deep Learning , 2017, NIPS.
[28] Tamás Sarlós,et al. Improved Approximation Algorithms for Large Matrices via Random Projections , 2006, 2006 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS'06).
[29] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[30] Ming Yang,et al. Compressing Deep Convolutional Networks using Vector Quantization , 2014, ArXiv.
[31] Pierre Baldi,et al. The dropout learning algorithm , 2014, Artif. Intell..
[32] Avner Magen,et al. Low rank matrix-valued chernoff bounds and approximate matrix multiplication , 2010, SODA '11.
[33] Konstantin Kutzkov,et al. Deterministic algorithms for skewed matrix products , 2012, STACS.
[34] Sagar V. Kamarthi,et al. Accelerating neural network training using weight extrapolations , 1999, Neural Networks.
[35] Petros Drineas,et al. Fast Monte-Carlo algorithms for approximate matrix multiplication , 2001, Proceedings 2001 IEEE International Conference on Cluster Computing.
[36] Xu Sun,et al. meProp: Sparsified Back Propagation for Accelerated Deep Learning with Reduced Overfitting , 2017, ICML.
[37] Sida I. Wang,et al. Dropout Training as Adaptive Regularization , 2013, NIPS.
[38] Joan Bruna,et al. Exploiting Linear Structure Within Convolutional Networks for Efficient Evaluation , 2014, NIPS.