Labeled Data Generation with Encoder-Decoder LSTM for Semantic Slot Filling
暂无分享,去创建一个
[1] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[2] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[3] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[4] Michael I. Jordan. Serial Order: A Parallel Distributed Processing Approach , 1997 .
[5] P. J. Price,et al. Evaluation of Spoken Language Systems: the ATIS Domain , 1990, HLT.
[6] Geoffrey Zweig,et al. Using Recurrent Neural Networks for Slot Filling in Spoken Language Understanding , 2015, IEEE/ACM Transactions on Audio, Speech, and Language Processing.
[7] Richard M. Schwartz,et al. Hidden Understanding Models of Natural Language , 1994, ACL.
[8] Wojciech Zaremba,et al. An Empirical Exploration of Recurrent Network Architectures , 2015, ICML.
[9] Phil Blunsom,et al. Recurrent Continuous Translation Models , 2013, EMNLP.
[10] Yoshua. Bengio,et al. Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..
[11] Andrew McCallum,et al. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.
[12] Dilek Z. Hakkani-Tür,et al. Spoken language understanding , 2008, IEEE Signal Processing Magazine.
[13] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[14] Gökhan Tür,et al. What is left to be understood in ATIS? , 2010, 2010 IEEE Spoken Language Technology Workshop.
[15] Steve Renals,et al. A study of the recurrent neural network encoder-decoder for large vocabulary speech recognition , 2015, INTERSPEECH.
[16] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[17] Geoffrey Zweig,et al. Sequence-to-sequence neural net models for grapheme-to-phoneme conversion , 2015, INTERSPEECH.
[18] Geoffrey Zweig,et al. Recurrent neural networks for language understanding , 2013, INTERSPEECH.
[19] Stephanie Seneff,et al. TINA: A Natural Language System for Spoken Language Applications , 1992, Comput. Linguistics.
[20] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.
[21] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[22] Alexander I. Rudnicky,et al. Expanding the Scope of the ATIS Task: The ATIS-3 Corpus , 1994, HLT.
[23] Yee Whye Teh,et al. A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.
[24] Daniel Jurafsky,et al. A Hierarchical Neural Autoencoder for Paragraphs and Documents , 2015, ACL.
[25] Sanjeev Khudanpur,et al. Variational approximation of long-span language models for lvcsr , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[26] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[27] Ngoc Thang Vu,et al. Bi-directional recurrent neural network with ranking loss for spoken language understanding , 2016, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[28] Alex Acero,et al. Combining Statistical and Knowledge-Based Spoken Language Understanding in Conditional Models , 2006, ACL.
[29] Yoshua Bengio,et al. Attention-Based Models for Speech Recognition , 2015, NIPS.
[30] Geoffrey Zweig,et al. Spoken language understanding using long short-term memory neural networks , 2014, 2014 IEEE Spoken Language Technology Workshop (SLT).
[31] Kaisheng Yao,et al. Recurrent Neural Networks with External Memory for Language Understanding , 2015, ArXiv.
[32] Jeffrey L. Elman,et al. Finding Structure in Time , 1990, Cogn. Sci..
[33] Yangyang Shi,et al. RNN-based labeled data generation for spoken language understanding , 2015, INTERSPEECH.