暂无分享,去创建一个
Yan Wang | Shen Furao | Jian Zhao | Hongyan Hao | Yudi Xia | Yan Wang | S. Furao | Hongyan Hao | Yudi Xia | Jian Zhao
[1] Yoshua Bengio,et al. Zoneout: Regularizing RNNs by Randomly Preserving Hidden Activations , 2016, ICLR.
[2] Guigang Zhang,et al. Deep Learning , 2016, Int. J. Semantic Comput..
[3] Daniel Jurafsky,et al. Sharp Nearby, Fuzzy Far Away: How Neural Language Models Use Context , 2018, ACL.
[4] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[5] Alex Graves,et al. Conditional Image Generation with PixelCNN Decoders , 2016, NIPS.
[6] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[7] Shuai Li,et al. Independently Recurrent Neural Network (IndRNN): Building A Longer and Deeper RNN , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[8] Richard Socher,et al. Pointer Sentinel Mixture Models , 2016, ICLR.
[9] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[10] Richard Socher,et al. Regularizing and Optimizing LSTM Language Models , 2017, ICLR.
[11] Heiga Zen,et al. WaveNet: A Generative Model for Raw Audio , 2016, SSW.
[12] Geoffrey E. Hinton,et al. Phoneme recognition using time-delay neural networks , 1989, IEEE Trans. Acoust. Speech Signal Process..
[13] Ruslan Salakhutdinov,et al. Breaking the Softmax Bottleneck: A High-Rank RNN Language Model , 2017, ICLR.
[14] Vladlen Koltun,et al. Trellis Networks for Sequence Modeling , 2018, ICLR.
[15] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[16] Peter A. Flach,et al. Proceedings of the 28th International Conference on Machine Learning , 2011 .
[17] Vladlen Koltun,et al. An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling , 2018, ArXiv.
[18] Dilin Wang,et al. Improving Neural Language Modeling via Adversarial Training , 2019, ICML.