暂无分享,去创建一个
Ran El-Yaniv | Yoshua Bengio | Daniel Soudry | Itay Hubara | Matthieu Courbariaux | Yoshua Bengio | Daniel Soudry | Itay Hubara | Matthieu Courbariaux | Ran El-Yaniv
[1] H. Spang,et al. Reduction of Quantizing Noise by Use of Feedback , 1962 .
[2] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[3] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[4] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[5] Viktor K. Prasanna,et al. Analysis of high-performance floating-point arithmetic on FPGAs , 2004, 18th International Parallel and Distributed Processing Symposium, 2004. Proceedings..
[6] Karl S. Hemmert,et al. Embedded floating-point units in FPGAs , 2006, FPGA '06.
[7] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[8] Berin Martini,et al. NeuFlow: A runtime reconfigurable dataflow processor for vision , 2011, CVPR 2011 WORKSHOPS.
[9] Clément Farabet,et al. Torch7: A Matlab-like Environment for Machine Learning , 2011, NIPS 2011.
[10] Alex Graves,et al. Practical Variational Inference for Neural Networks , 2011, NIPS.
[11] Yoshua Bengio,et al. Domain Adaptation for Large-Scale Sentiment Classification: A Deep Learning Approach , 2011, ICML.
[12] John Langford,et al. Scaling up machine learning: parallel and distributed approaches , 2011, KDD '11 Tutorials.
[13] Berin Martini,et al. Large-Scale FPGA-based Convolutional Networks , 2011 .
[14] Vincent Vanhoucke,et al. Improving the speed of neural networks on CPUs , 2011 .
[15] Razvan Pascanu,et al. Theano: new features and speed improvements , 2012, ArXiv.
[16] E. Culurciello,et al. NeuFlow: Dataflow vision processing system-on-a-chip , 2012, 2012 IEEE 55th International Midwest Symposium on Circuits and Systems (MWSCAS).
[17] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[18] Geoffrey Zweig,et al. Context dependent recurrent neural network language model , 2012, 2012 IEEE Spoken Language Technology Workshop (SLT).
[19] Tara N. Sainath,et al. Deep Neural Networks for Acoustic Modeling in Speech Recognition , 2012 .
[20] Yoshua Bengio,et al. Estimating or Propagating Gradients Through Stochastic Neurons , 2013, ArXiv.
[21] Yann LeCun,et al. Regularization of Neural Networks using DropConnect , 2013, ICML.
[22] Yichuan Tang,et al. Deep Learning using Linear Support Vector Machines , 2013, 1306.0239.
[23] Tara N. Sainath,et al. Deep convolutional neural networks for LVCSR , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[24] Ian J. Goodfellow,et al. Pylearn2: a machine learning research library , 2013, ArXiv.
[25] Yoshua Bengio,et al. Maxout Networks , 2013, ICML.
[26] Tao Wang,et al. Deep learning with COTS HPC systems , 2013, ICML.
[27] Richard M. Schwartz,et al. Fast and Robust Neural Network Joint Models for Statistical Machine Translation , 2014, ACL.
[28] Yoshua Bengio,et al. Training deep neural networks with low precision multiplications , 2014 .
[29] Ninghui Sun,et al. DianNao: a small-footprint high-throughput accelerator for ubiquitous machine-learning , 2014, ASPLOS.
[30] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[31] Jia Wang,et al. DaDianNao: A Machine-Learning Supercomputer , 2014, 2014 47th Annual IEEE/ACM International Symposium on Microarchitecture.
[32] Parul Parashar,et al. Neural Networks in Machine Learning , 2014 .
[33] Alex Krizhevsky,et al. One weird trick for parallelizing convolutional neural networks , 2014, ArXiv.
[34] Mark Horowitz,et al. 1.1 Computing's energy problem (and what we can do about it) , 2014, 2014 IEEE International Solid-State Circuits Conference Digest of Technical Papers (ISSCC).
[35] Kyuyeon Hwang,et al. Fixed-point feedforward deep neural network design using weights +1, 0, and −1 , 2014, 2014 IEEE Workshop on Signal Processing Systems (SiPS).
[36] Benjamin Graham,et al. Spatially-sparse convolutional neural networks , 2014, ArXiv.
[37] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[38] Ming Yang,et al. Compressing Deep Convolutional Networks using Vector Quantization , 2014, ArXiv.
[39] Ron Meir,et al. Expectation Backpropagation: Parameter-Free Training of Multilayer Neural Networks with Continuous or Discrete Weights , 2014, NIPS.
[40] Dharmendra S. Modha,et al. Backpropagation for Energy-Efficient Neuromorphic Computing , 2015, NIPS.
[41] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[42] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[43] Carlo Baldassi,et al. Subdominant Dense Clusters Allow for Simple Learning and High Computational Performance in Neural Networks with Discrete Synapses. , 2015, Physical review letters.
[44] Colin Raffel,et al. Lasagne: First release. , 2015 .
[45] Daniel Soudry,et al. Training Binary Multilayer Neural Networks for Image Classification using Expectation Backpropagation , 2015, ArXiv.
[46] Yoshua Bengio,et al. BinaryConnect: Training Deep Neural Networks with binary weights during propagations , 2015, NIPS.
[47] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[48] Alexander Mordvintsev,et al. Inceptionism: Going Deeper into Neural Networks , 2015 .
[49] Jian Sun,et al. Efficient and accurate approximations of nonlinear convolutional networks , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[50] Pritish Narayanan,et al. Deep Learning with Limited Numerical Precision , 2015, ICML.
[51] Yoshua Bengio,et al. FitNets: Hints for Thin Deep Nets , 2014, ICLR.
[52] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[53] Dumitru Erhan,et al. Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[54] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[55] Yixin Chen,et al. Compressing Neural Networks with the Hashing Trick , 2015, ICML.
[56] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[57] Zhuowen Tu,et al. Deeply-Supervised Nets , 2014, AISTATS.
[58] Luca Benini,et al. YodaNN: An Ultra-Low Power Convolutional Neural Network Accelerator Based on Binary Weights , 2016, 2016 IEEE Computer Society Annual Symposium on VLSI (ISVLSI).
[59] Dharmendra S. Modha,et al. Deep neural networks are robust to weight binarization and other non-linear distortions , 2016, ArXiv.
[60] Zhuowen Tu,et al. Generalizing Pooling Functions in Convolutional Neural Networks: Mixed, Gated, and Tree , 2015, AISTATS.
[61] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[62] Yoshua Bengio,et al. Neural Networks with Few Multiplications , 2015, ICLR.
[63] Ying Zhang,et al. Recurrent Neural Networks With Limited Numerical Precision , 2016, ArXiv.
[64] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[65] Shuchang Zhou,et al. DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients , 2016, ArXiv.
[66] Paris Smaragdis,et al. Bitwise Neural Networks , 2016, ArXiv.
[67] Tsutomu Matsumoto,et al. ASIC implementation of random number generators using SR latches and its evaluation , 2016, EURASIP J. Inf. Secur..
[68] Soheil Ghiasi,et al. Hardware-oriented Approximation of Convolutional Neural Networks , 2016, ArXiv.
[69] Igor Carron,et al. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks , 2016 .
[70] Daisuke Miyashita,et al. Convolutional Neural Networks using Logarithmic Data Representation , 2016, ArXiv.
[71] Luca Benini,et al. YodaNN: An Architecture for Ultralow Power Binary-Weight CNN Acceleration , 2016, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.