Efficient Semi-Supervised Learning for Natural Language Understanding by Optimizing Diversity
暂无分享,去创建一个
William M. Campbell | John P. Lalor | Eunah Cho | Varun Kumar | He Xie | Eunah Cho | W. Campbell | He Xie | Varun Kumar
[1] Anima Anandkumar,et al. Deep Active Learning for Named Entity Recognition , 2017, Rep4NLP@ACL.
[2] Kevin Gimpel,et al. Towards Universal Paraphrastic Sentence Embeddings , 2015, ICLR.
[3] Frédéric Béchet,et al. Online SLU model adaptation with a partial oracle , 2010, INTERSPEECH.
[4] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[5] Philipp Koehn,et al. Statistical Significance Tests for Machine Translation Evaluation , 2004, EMNLP.
[6] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[7] M. Friedman. A Comparison of Alternative Tests of Significance for the Problem of $m$ Rankings , 1940 .
[8] Feng Gao,et al. Spoken language understanding using weakly supervised learning , 2010, Comput. Speech Lang..
[9] Ralph Weischedel,et al. PERFORMANCE MEASURES FOR INFORMATION EXTRACTION , 2007 .
[10] Gökhan Tür,et al. Combining active and semi-supervised learning for spoken language understanding , 2005, Speech Commun..
[11] Alexander Zien,et al. Semi-Supervised Learning , 2006 .
[12] Xiaojin Zhu,et al. --1 CONTENTS , 2006 .
[13] Matteo Pagliardini,et al. Unsupervised Learning of Sentence Embeddings Using Compositional n-Gram Features , 2017, NAACL.
[14] Gokhan Tur,et al. Spoken Language Understanding: Systems for Extracting Semantic Information from Speech , 2011 .
[15] Eunah Cho,et al. Paraphrase Generation for Semi-Supervised Learning in NLU , 2019, Proceedings of the Workshop on Methods for Optimizing and Evaluating Neural Language Generation.
[16] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[17] Richard M. Schwartz,et al. Unsupervised Training on Large Amounts of Broadcast News Data , 2006, 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings.
[18] Rahul Gupta,et al. A Re-Ranker Scheme For Integrating Large Scale NLU Models , 2018, 2018 IEEE Spoken Language Technology Workshop (SLT).
[19] Andreas Krause,et al. Cost-effective outbreak detection in networks , 2007, KDD '07.
[20] Colin Raffel,et al. Realistic Evaluation of Deep Semi-Supervised Learning Algorithms , 2018, NeurIPS.
[21] Hal Daumé,et al. Deep Unordered Composition Rivals Syntactic Methods for Text Classification , 2015, ACL.
[22] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[23] Ari Rappoport,et al. Self-Training for Enhancement and Domain Adaptation of Statistical Parsers Trained on Small Datasets , 2007, ACL.
[24] Michel Minoux,et al. Accelerated greedy algorithms for maximizing submodular set functions , 1978 .
[25] Xiaojin Zhu,et al. Introduction to Semi-Supervised Learning , 2009, Synthesis Lectures on Artificial Intelligence and Machine Learning.
[26] Jeff A. Bilmes,et al. Submodularity for Data Selection in Machine Translation , 2014, EMNLP.
[27] Silvio Savarese,et al. Active Learning for Convolutional Neural Networks: A Core-Set Approach , 2017, ICLR.
[28] R. E. Lee,et al. Distribution-free multiple comparisons between successive treatments , 1995 .
[29] Eugene Charniak,et al. Effective Self-Training for Parsing , 2006, NAACL.
[30] Jason Weston,et al. A unified architecture for natural language processing: deep neural networks with multitask learning , 2008, ICML '08.
[31] Chris Callison-Burch,et al. PPDB 2.0: Better paraphrase ranking, fine-grained entailment relations, word embeddings, and style classification , 2015, ACL.
[32] Chu-Ren Huang,et al. Semi-supervised Text Categorization by Considering Sufficiency and Diversity , 2013, NLPCC.