Deep Cascade Multi-Task Learning for Slot Filling in Online Shopping Assistant
暂无分享,去创建一个
Muhua Zhu | Xi Chen | Kenny Q. Zhu | Wenwu Ou | Yu Gong | Zhao Li | Lu Duan | Xusheng Luo | Yu Zhu | Z. Li | Muhua Zhu | Yu Zhu | Wenwu Ou | Xusheng Luo | Yu Gong | Xi Chen | Lu Duan | Zhao Li
[1] Rich Caruana,et al. Multitask Learning , 1998, Encyclopedia of Machine Learning and Data Mining.
[2] Iryna Gurevych,et al. Optimal Hyperparameters for Deep LSTM-Networks for Sequence Labeling Tasks , 2017, ArXiv.
[3] Nanyun Peng,et al. Multi-task Multi-domain Representation Learning for Sequence Tagging , 2016, ArXiv.
[4] Geoffrey Zweig,et al. Spoken language understanding using long short-term memory neural networks , 2014, 2014 IEEE Spoken Language Technology Workshop (SLT).
[5] Deng Cai,et al. What to Do Next: Modeling User Behaviors by Time-LSTM , 2017, IJCAI.
[6] Anders Søgaard,et al. Deep multi-task learning with low level tasks supervised at lower layers , 2016, ACL.
[7] Bowen Zhou,et al. Neural Models for Sequence Chunking , 2017, AAAI.
[8] Nanyun Peng,et al. Improving Named Entity Recognition for Chinese Social Media with Word Segmentation Representation Learning , 2016, ACL.
[9] Gökhan Tür,et al. Multi-Domain Joint Semantic Frame Parsing Using Bi-Directional RNN-LSTM , 2016, INTERSPEECH.
[10] Shih-Fu Chang,et al. Deep Cross Residual Learning for Multitask Visual Recognition , 2016, ACM Multimedia.
[11] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Ngoc Thang Vu,et al. Bi-directional recurrent neural network with ranking loss for spoken language understanding , 2016, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[13] Ruslan Salakhutdinov,et al. Transfer Learning for Sequence Tagging with Hierarchical Recurrent Networks , 2016, ICLR.
[14] Bing Liu,et al. Joint Online Spoken Language Understanding and Language Modeling With Recurrent Neural Networks , 2016, SIGDIAL Conference.
[15] Deng Cai,et al. A Brand-level Ranking System with the Customized Attention-GRU Model , 2018, IJCAI.
[16] Geoffrey Zweig,et al. Using Recurrent Neural Networks for Slot Filling in Spoken Language Understanding , 2015, IEEE/ACM Transactions on Audio, Speech, and Language Processing.
[17] BengioYoshua,et al. Using recurrent neural networks for slot filling in spoken language understanding , 2015 .
[18] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[19] Giuseppe Riccardi,et al. Generative and discriminative algorithms for spoken language understanding , 2007, INTERSPEECH.
[20] P. J. Price,et al. Evaluation of Spoken Language Systems: the ATIS Domain , 1990, HLT.
[21] Geoffrey Zweig,et al. Recurrent neural networks for language understanding , 2013, INTERSPEECH.
[22] Kai Yu,et al. Encoder-decoder with focus-mechanism for sequence labelling based spoken language understanding , 2016, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[23] Steve Young,et al. A data-driven spoken language understanding system , 2003, 2003 IEEE Workshop on Automatic Speech Recognition and Understanding (IEEE Cat. No.03EX721).
[24] Ye-Yi Wang,et al. Spoken language understanding , 2005, IEEE Signal Processing Magazine.
[25] Zhoujun Li,et al. Building Task-Oriented Dialogue Systems for Online Shopping , 2017, AAAI.
[26] Bing Liu,et al. Recurrent Neural Network Structured Output Prediction for Spoken Language Understanding , 2015 .
[27] Ngoc Thang Vu. Sequential Convolutional Neural Networks for Slot Filling in Spoken Language Understanding , 2016, INTERSPEECH.
[28] Wei Xu,et al. Bidirectional LSTM-CRF Models for Sequence Tagging , 2015, ArXiv.
[29] Ruhi Sarikaya,et al. Convolutional neural network based triangular CRF for joint intent detection and slot filling , 2013, 2013 IEEE Workshop on Automatic Speech Recognition and Understanding.
[30] Dilek Z. Hakkani-Tür,et al. Spoken language understanding , 2008, IEEE Signal Processing Magazine.
[31] Yuji Matsumoto,et al. Chunking with Support Vector Machines , 2001, NAACL.
[32] Samy Bengio,et al. Show and tell: A neural image caption generator , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[33] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[34] Houfeng Wang,et al. A Joint Model of Intent Determination and Slot Filling for Spoken Language Understanding , 2016, IJCAI.
[35] Daniel Jurafsky,et al. Distant supervision for relation extraction without labeled data , 2009, ACL.
[36] Andrew McCallum,et al. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.
[37] Frédéric Béchet,et al. Is ATIS Too Shallow to Go Deeper for Benchmarking Spoken Language Understanding Models? , 2018, INTERSPEECH.
[38] Bowen Zhou,et al. Leveraging Sentence-level Information with Encoder LSTM for Semantic Slot Filling , 2016, EMNLP.
[39] Kaisheng Yao,et al. Recurrent Neural Networks with External Memory for Language Understanding , 2015, ArXiv.