Pseudo Labeling and Negative Feedback Learning for Large-Scale Multi-Label Domain Classification
暂无分享,去创建一个
[1] Tatsuya Harada,et al. Multi-label Ranking from Positive and Unlabeled Data , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[2] Eric P. Xing,et al. Harnessing Deep Neural Networks with Logic Rules , 2016, ACL.
[3] Gokhan Tur,et al. Spoken Language Understanding: Systems for Extracting Semantic Information from Speech , 2011 .
[4] Barbara Plank,et al. Strong Baselines for Neural Semi-Supervised Learning under Domain Shift , 2018, ACL.
[5] Andrew McCallum,et al. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.
[6] Young-Bum Kim,et al. Supervised Domain Enablement Attention for Personalized Domain Classification , 2018, EMNLP.
[7] Young-Bum Kim,et al. An overview of end-to-end language understanding and dialog management for personal digital assistants , 2016, 2016 IEEE Spoken Language Technology Workshop (SLT).
[8] Inderjit S. Dhillon,et al. Large-scale Multi-label Learning with Missing Labels , 2013, ICML.
[9] Jason Weston,et al. Learning from Dialogue after Deployment: Feed Yourself, Chatbot! , 2019, ACL.
[10] Yoon Kim,et al. Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.
[11] Jürgen Schmidhuber,et al. Framewise phoneme classification with bidirectional LSTM and other neural network architectures , 2005, Neural Networks.
[12] Björn Hoffmeister,et al. Just ASK: Building an Architecture for Extensible Self-Service Spoken Language Understanding , 2017, ArXiv.
[13] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[14] Dong-Hyun Lee,et al. Pseudo-Label : The Simple and Efficient Semi-Supervised Learning Method for Deep Neural Networks , 2013 .
[15] Young-Bum Kim,et al. A Scalable Neural Shortlisting-Reranking Approach for Large-Scale Domain Classification in Natural Language Understanding , 2018, NAACL.
[16] Harri Valpola,et al. Weight-averaged consistency targets improve semi-supervised deep learning results , 2017, ArXiv.
[17] Zachary Chase Lipton,et al. Born Again Neural Networks , 2018, ICML.
[18] Gang Niu,et al. Classification from Positive, Unlabeled and Biased Negative Data , 2018, ICML.
[19] David Yarowsky,et al. Unsupervised Word Sense Disambiguation Rivaling Supervised Methods , 1995, ACL.
[20] Ruhi Sarikaya,et al. Hypotheses ranking for robust domain classification and tracking in dialogue systems , 2014, INTERSPEECH.
[21] Yoshua Bengio,et al. Semi-supervised Learning by Entropy Minimization , 2004, CAP.
[22] Jason Weston,et al. Dialog-based Language Learning , 2016, NIPS.
[23] Bowen Zhou,et al. A Structured Self-attentive Sentence Embedding , 2017, ICLR.
[24] Eugene Charniak,et al. Reranking and Self-Training for Parser Adaptation , 2006, ACL.
[25] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.