Attention-Based CNN-BLSTM Networks for Joint Intent Detection and Slot Filling
暂无分享,去创建一个
[1] Sebastian Weigelt,et al. Integrating a Dialog Component into a Framework for Spoken Language Understanding , 2018, 2018 IEEE/ACM 6th International Workshop on Realizing Artificial Intelligence Synergies in Software Engineering (RAISE).
[2] Geoffrey E. Hinton,et al. Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[3] George R. Doddington,et al. The ATIS Spoken Language Systems Pilot Corpus , 1990, HLT.
[4] Zhiyuan Liu,et al. A C-LSTM Neural Network for Text Classification , 2015, ArXiv.
[5] Bing Liu,et al. Attention-Based Recurrent Neural Network Models for Joint Intent Detection and Slot Filling , 2016, INTERSPEECH.
[6] Geoffrey Zweig,et al. Joint semantic utterance classification and slot filling with recursive neural networks , 2014, 2014 IEEE Spoken Language Technology Workshop (SLT).
[7] Houkuan Huang,et al. Feature selection for text classification with Naïve Bayes , 2009, Expert Syst. Appl..
[8] Bowen Zhou,et al. Leveraging Sentence-level Information with Encoder LSTM for Natural Language Understanding , 2016, ArXiv.
[9] Ruhi Sarikaya,et al. Convolutional neural network based triangular CRF for joint intent detection and slot filling , 2013, 2013 IEEE Workshop on Automatic Speech Recognition and Understanding.
[10] Yoshua Bengio,et al. Hierarchical Probabilistic Neural Network Language Model , 2005, AISTATS.
[11] Gökhan Tür,et al. Optimizing SVMs for complex call classification , 2003, 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing, 2003. Proceedings. (ICASSP '03)..
[12] Yoon Kim,et al. Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.
[13] Geoffrey Zweig,et al. Spoken language understanding using long short-term memory neural networks , 2014, 2014 IEEE Spoken Language Technology Workshop (SLT).
[14] Bowen Zhou,et al. ABCNN: Attention-Based Convolutional Neural Network for Modeling Sentence Pairs , 2015, TACL.
[15] Wojciech Zaremba,et al. An Empirical Exploration of Recurrent Network Architectures , 2015, ICML.
[16] Bing Liu,et al. Joint Online Spoken Language Understanding and Language Modeling With Recurrent Neural Networks , 2016, SIGDIAL Conference.
[17] Shahram Khadivi,et al. Graph-Based Semi-Supervised Conditional Random Fields For Spoken Language Understanding Using Unaligned Data , 2014, ALTA.
[18] Bowen Zhou,et al. Leveraging Sentence-level Information with Encoder LSTM for Semantic Slot Filling , 2016, EMNLP.
[19] Xiaolong Wang,et al. The study of a nonstationary maximum entropy Markov model and its application on the pos-tagging task , 2007, TALIP.
[20] Giuseppe Riccardi,et al. Generative and discriminative algorithms for spoken language understanding , 2007, INTERSPEECH.
[21] Ngoc Thang Vu,et al. Bi-directional recurrent neural network with ranking loss for spoken language understanding , 2016, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[22] Kai Yu,et al. Encoder-decoder with focus-mechanism for sequence labelling based spoken language understanding , 2016, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[23] Diana Inkpen,et al. Speech Intent Recognition for Robots , 2016, 2016 Third International Conference on Mathematics and Computers in Sciences and in Industry (MCSI).
[24] Houfeng Wang,et al. A Joint Model of Intent Determination and Slot Filling for Spoken Language Understanding , 2016, IJCAI.
[25] Navdeep Jaitly,et al. Hybrid speech recognition with Deep Bidirectional LSTM , 2013, 2013 IEEE Workshop on Automatic Speech Recognition and Understanding.
[26] Kyunghyun Cho,et al. Efficient Character-level Document Classification by Combining Convolution and Recurrent Layers , 2016, ArXiv.