Optimizing Deep Learning RNN Topologies on Intel Architecture
暂无分享,去创建一个
Alexander Heinecke | Kunal Banerjee | Evangelos Georganas | Dhiraj D. Kalamkar | Eden Segal | Cristina S. Anderson | Barukh Ziv | E. Georganas | Barukh Ziv | A. Heinecke | Eden Segal | K. Banerjee
[1] David Vandyke,et al. Semantically Conditioned LSTM-based Natural Language Generation for Spoken Dialogue Systems , 2015, EMNLP.
[2] M. Powell,et al. Approximation theory and methods , 1984 .
[3] Majid Ahmadi,et al. Efficient hardware implementation of the hyperbolic tangent sigmoid function , 2009, 2009 IEEE International Symposium on Circuits and Systems.
[4] Matti Tommiska,et al. Efficient digital implementation of the sigmoid function for reprogrammable logic , 2003 .
[5] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[6] Alexander Heinecke,et al. Training Google Neural Machine Translation on an Intel CPU Cluster , 2019, 2019 IEEE International Conference on Cluster Computing (CLUSTER).
[7] William James Cody,et al. Software Manual for the Elementary Functions (Prentice-Hall series in computational mathematics) , 1980 .
[8] Yuan Yu,et al. TensorFlow: A system for large-scale machine learning , 2016, OSDI.
[9] Qiang Liu,et al. Reference Based LSTM for Image Captioning , 2017, AAAI.
[10] C. Brezinski. Outlines of Pade Approximation , 1983 .
[11] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[12] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[13] John Tran,et al. cuDNN: Efficient Primitives for Deep Learning , 2014, ArXiv.
[14] Yoshua Bengio,et al. Training deep neural networks with low precision multiplications , 2014 .
[15] Pradeep Dubey,et al. Mixed Precision Training of Convolutional Neural Networks using Integer Operations , 2018, ICLR.