Wavefront parallelization of recurrent neural networks on multi-core architectures
暂无分享,去创建一个
[1] Mike Schuster,et al. On supervised learning from sequential data with applications for speech regognition , 1999 .
[2] Honglak Lee,et al. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations , 2009, ICML '09.
[3] David A. Padua,et al. Advanced compiler optimizations for supercomputers , 1986, CACM.
[4] Jonathan Schaeffer,et al. Generating parallel programs from the wavefront design pattern , 2002, Proceedings 16th International Parallel and Distributed Processing Symposium.
[5] Jian Zhang,et al. Exploiting Model-Level Parallelism in Recurrent Neural Network Accelerators , 2019, 2019 IEEE 13th International Symposium on Embedded Multicore/Many-core Systems-on-Chip (MCSoC).
[6] Jürgen Schmidhuber,et al. LSTM recurrent networks learn simple context-free and context-sensitive languages , 2001, IEEE Trans. Neural Networks.
[7] Guigang Zhang,et al. Deep Learning , 2016, Int. J. Semantic Comput..
[8] Geoffrey E. Hinton,et al. Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[9] Jack B. Dennis,et al. Data Flow Supercomputers , 1980, Computer.
[10] Anthony J. Robinson,et al. Static and Dynamic Error Propagation Networks with Application to Speech Coding , 1987, NIPS.
[11] Yoshua Bengio,et al. On the Properties of Neural Machine Translation: Encoder–Decoder Approaches , 2014, SSST@EMNLP.
[12] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[13] Yu Zhang,et al. Simple Recurrent Units for Highly Parallelizable Recurrence , 2017, EMNLP.
[14] R. Karp,et al. Properties of a model for parallel computations: determinacy , 1966 .
[15] Yoshua Bengio,et al. Exploring Strategies for Training Deep Neural Networks , 2009, J. Mach. Learn. Res..
[16] Stephen Marshall,et al. Activation Functions: Comparison of trends in Practice and Research for Deep Learning , 2018, ArXiv.
[17] James C. Christensen,et al. Deep long short-term memory structures model temporal dependencies improving cognitive workload estimation , 2017, Pattern Recognit. Lett..
[18] L. Dagum,et al. OpenMP: an industry standard API for shared-memory programming , 1998 .
[19] Trevor Darrell,et al. Caffe: Convolutional Architecture for Fast Feature Embedding , 2014, ACM Multimedia.
[20] Phil Blunsom,et al. Optimizing Performance of Recurrent Neural Networks on GPUs , 2016, ArXiv.
[21] Shih-Chii Liu,et al. Phased LSTM: Accelerating Recurrent Network Training for Long or Event-based Sequences , 2016, NIPS.
[22] Yaroslav Bulatov,et al. Multi-digit Number Recognition from Street View Imagery using Deep Convolutional Neural Networks , 2013, ICLR.
[23] Alexander J. Smola,et al. Efficient mini-batch training for stochastic optimization , 2014, KDD.
[24] Yoshua Bengio,et al. Light Gated Recurrent Units for Speech Recognition , 2018, IEEE Transactions on Emerging Topics in Computational Intelligence.
[25] Rafael Asenjo,et al. A case study of the task-based parallel wavefront pattern , 2011, PARCO.
[26] Karl J. Ottenstein,et al. The program dependence graph in a software development environment , 1984, SDE 1.
[27] Klaus Nordhausen,et al. The Elements of Statistical Learning: Data Mining, Inference, and Prediction, Second Edition by Trevor Hastie, Robert Tibshirani, Jerome Friedman , 2009 .
[28] Alejandro Duran,et al. Ompss: a Proposal for Programming Heterogeneous Multi-Core Architectures , 2011, Parallel Process. Lett..
[29] Benjamin Schrauwen,et al. Training and analyzing deep recurrent neural networks , 2013, NIPS 2013.
[30] Shahrokh Valaee,et al. Interpretation of Mammogram and Chest Radiograph Reports using Deep Neural Networks - Preliminary Results , 2017, ArXiv.
[31] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.
[32] Andrew W. Senior,et al. Long short-term memory recurrent neural network architectures for large scale acoustic modeling , 2014, INTERSPEECH.
[33] Stefan Braun,et al. LSTM Benchmarks for Deep Learning Frameworks , 2018, ArXiv.
[34] Haibin Kan,et al. Cache-oblivious wavefront: improving parallelism of recursive dynamic programming algorithms without losing cache-efficiency , 2015, PPoPP.
[35] Geoffrey E. Hinton,et al. Generating Text with Recurrent Neural Networks , 2011, ICML.
[36] Lukasz Kaiser,et al. Neural GPUs Learn Algorithms , 2015, ICLR.
[37] Zheng Zhang,et al. MXNet: A Flexible and Efficient Machine Learning Library for Heterogeneous Distributed Systems , 2015, ArXiv.
[38] M. Georgiopoulos,et al. Feed-forward neural networks , 1994, IEEE Potentials.
[39] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[40] Minjia Zhang,et al. DeepCPU: Serving RNN-based Deep Learning Models 10x Faster , 2018, USENIX Annual Technical Conference.
[41] Lawrence Snyder,et al. Pipelining Wavefront Computations: Experiences and Performance , 2000, IPDPS Workshops.
[42] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[43] Marcus Liwicki,et al. A novel approach to on-line handwriting recognition based on bidirectional long short-term memory networks , 2007 .
[44] J OttensteinKarl,et al. The program dependence graph in a software development environment , 1984 .