暂无分享,去创建一个
Reza Yazdani | Olatunji Ruwase | Minjia Zhang | Yuxiong He | Jose-Maria Arnau | Antonio Gonzalez | Olatunji Ruwase | Yuxiong He | Minjia Zhang | Antonio González | R. Yazdani | José-María Arnau | J. Arnau
[1] Björn W. Schuller,et al. Introducing CURRENNT: the munich open-source CUDA recurrent neural network toolkit , 2015, J. Mach. Learn. Res..
[2] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[3] Shahrokh Valaee,et al. Recent Advances in Recurrent Neural Networks , 2017, ArXiv.
[4] Berin Martini,et al. Recurrent Neural Networks Hardware Implementation on FPGA , 2015, ArXiv.
[5] João Canas Ferreira,et al. An FPGA implementation of a long short-term memory neural network , 2016, 2016 International Conference on ReConFigurable Computing and FPGAs (ReConFig).
[6] J. A. Hartigan,et al. A k-means clustering algorithm , 1979 .
[7] Pedro López,et al. Design of Hybrid Second-Level Caches , 2015, IEEE Transactions on Computers.
[8] Michael Hübner,et al. A Survey on CNN and RNN Implementations , 2017 .
[9] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[10] Jason Cong,et al. FPGA-based accelerator for long short-term memory recurrent neural networks , 2017, 2017 22nd Asia and South Pacific Design Automation Conference (ASP-DAC).
[11] Song Han,et al. ESE: Efficient Speech Recognition Engine with Compressed LSTM on FPGA , 2016, ArXiv.
[12] Quoc V. Le,et al. SpecAugment: A Simple Data Augmentation Method for Automatic Speech Recognition , 2019, INTERSPEECH.
[13] Minjia Zhang,et al. DeepCPU: Serving RNN-based Deep Learning Models 10x Faster , 2018, USENIX Annual Technical Conference.
[14] Shuohang Wang,et al. An LSTM model for cloze-style machine comprehension , 2018 .
[15] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[16] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[17] Jung Ho Ahn,et al. CACTI-P: Architecture-level modeling for SRAM-based structures with advanced leakage reduction techniques , 2011, 2011 IEEE/ACM International Conference on Computer-Aided Design (ICCAD).
[18] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[19] Jose-Maria Arnau,et al. E-PUR: an energy-efficient processing unit for recurrent neural networks , 2017, PACT.
[20] Marco Baroni,et al. The emergence of number and syntax units in LSTM language models , 2019, NAACL.
[21] Jianwei Yu,et al. Gaussian Process Lstm Recurrent Neural Network Language Models for Speech Recognition , 2019, ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[22] Eric S. Chung,et al. A Configurable Cloud-Scale DNN Processor for Real-Time AI , 2018, 2018 ACM/IEEE 45th Annual International Symposium on Computer Architecture (ISCA).
[23] Erich Elsen,et al. Persistent RNNs: Stashing Recurrent Weights On-Chip , 2016, ICML.
[24] Trevor Darrell,et al. Long-term recurrent convolutional networks for visual recognition and description , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Ali Farhadi,et al. Bidirectional Attention Flow for Machine Comprehension , 2016, ICLR.
[26] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[27] Samy Bengio,et al. Show and tell: A neural image caption generator , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[28] Yajie Miao,et al. EESEN: End-to-end speech recognition using deep RNN models and WFST-based decoding , 2015, 2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU).
[29] Yuxiong He,et al. GRNN: Low-Latency and Scalable RNN Inference on GPUs , 2019, EuroSys.
[30] Qinru Qiu,et al. C-LSTM: Enabling Efficient LSTM using Structured Compression Techniques on FPGAs , 2018, FPGA.
[31] Yu Wang,et al. FPGA Acceleration of Recurrent Neural Network Based Language Model , 2015, 2015 IEEE 23rd Annual International Symposium on Field-Programmable Custom Computing Machines.
[32] John Tran,et al. cuDNN: Efficient Primitives for Deep Learning , 2014, ArXiv.
[33] Phil Blunsom,et al. Optimizing Performance of Recurrent Neural Networks on GPUs , 2016, ArXiv.
[34] Trevor Darrell,et al. Sequence to Sequence -- Video to Text , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[35] Song Han,et al. EIE: Efficient Inference Engine on Compressed Deep Neural Network , 2016, 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA).
[36] Yoshua Bengio,et al. Towards Non-saturating Recurrent Units for Modelling Long-term Dependencies , 2019, AAAI.
[37] Furu Wei,et al. Learning to Ask Unanswerable Questions for Machine Reading Comprehension , 2019, ACL.
[38] David A. Patterson,et al. In-datacenter performance analysis of a tensor processing unit , 2017, 2017 ACM/IEEE 44th Annual International Symposium on Computer Architecture (ISCA).
[39] Song Han,et al. From model to FPGA: Software-hardware co-design for efficient neural network acceleration , 2016, 2016 IEEE Hot Chips 28 Symposium (HCS).
[40] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.