暂无分享,去创建一个
Hannaneh Hajishirzi | Sachin Mehta | Rik Koncel-Kedziorski | Mohammad Rastegari | Mohammad Rastegari | Hannaneh Hajishirzi | Rik Koncel-Kedziorski | Sachin Mehta
[1] Ruslan Salakhutdinov,et al. Breaking the Softmax Bottleneck: A High-Rank RNN Language Model , 2017, ICLR.
[2] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[3] Richard Socher,et al. Pointer Sentinel Mixture Models , 2016, ICLR.
[4] M. Gevrey,et al. Review and comparison of methods to study the contribution of variables in artificial neural network models , 2003 .
[5] Ali Farhadi,et al. Neural Speed Reading via Skim-RNN , 2017, ICLR.
[6] Boris Ginsburg,et al. Factorization tricks for LSTM networks , 2017, ICLR.
[7] Linda G. Shapiro,et al. ESPNet: Efficient Spatial Pyramid of Dilated Convolutions for Semantic Segmentation , 2018, ECCV.
[8] Ali Farhadi,et al. Query-Reduction Networks for Question Answering , 2016, ICLR.
[9] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[10] Jürgen Schmidhuber,et al. Recurrent Highway Networks , 2016, ICML.
[11] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[12] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[13] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[14] Ariel D. Procaccia,et al. Variational Dropout and the Local Reparameterization Trick , 2015, NIPS.
[15] Jordi Torres,et al. Skip RNN: Learning to Skip State Updates in Recurrent Neural Networks , 2017, ICLR.
[16] Xinlei Chen,et al. Visualizing and Understanding Neural Models in NLP , 2015, NAACL.
[17] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[18] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[19] Hakan Inan,et al. Tying Word Vectors and Word Classifiers: A Loss Framework for Language Modeling , 2016, ICLR.
[20] Edward H. Adelson,et al. The Laplacian Pyramid as a Compact Image Code , 1983, IEEE Trans. Commun..
[21] Alexander M. Rush,et al. Character-Aware Neural Language Models , 2015, AAAI.
[22] Richard Socher,et al. Quasi-Recurrent Neural Networks , 2016, ICLR.
[23] Yoshua Bengio,et al. Convolutional networks for images, speech, and time series , 1998 .
[24] Junmo Kim,et al. Deep Pyramidal Residual Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Quoc V. Le,et al. Neural Architecture Search with Reinforcement Learning , 2016, ICLR.
[26] Jürgen Schmidhuber,et al. Recurrent nets that time and count , 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks. IJCNN 2000. Neural Computing: New Challenges and Perspectives for the New Millennium.
[27] David G. Lowe,et al. Object recognition from local scale-invariant features , 1999, Proceedings of the Seventh IEEE International Conference on Computer Vision.
[28] Quoc V. Le,et al. Learning to Skim Text , 2017, ACL.
[29] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[30] Omer Levy,et al. Recurrent Additive Networks , 2017, ArXiv.
[31] Richard Socher,et al. Regularizing and Optimizing LSTM Language Models , 2017, ICLR.
[32] Ben Krause,et al. Explorer Dynamic Evaluation of Neural Sequence Models , 2018 .
[33] Zoubin Ghahramani,et al. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.
[34] Klaus-Robert Müller,et al. Explaining Recurrent Neural Network Predictions in Sentiment Analysis , 2017, WASSA@EMNLP.
[35] Chris Dyer,et al. On the State of the Art of Evaluation in Neural Language Models , 2017, ICLR.
[36] Hongbin Zha,et al. Alternating Multi-bit Quantization for Recurrent Neural Networks , 2018, ICLR.
[37] Yu Zhang,et al. Training RNNs as Fast as CNNs , 2017, EMNLP 2018.