暂无分享,去创建一个
[1] John Blitzer,et al. Domain Adaptation with Structural Correspondence Learning , 2006, EMNLP.
[2] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[3] Christopher Potts,et al. A large annotated corpus for learning natural language inference , 2015, EMNLP.
[4] Qiang Yang,et al. Cross-domain sentiment classification via spectral feature alignment , 2010, WWW '10.
[5] Fan-Keng Sun,et al. Conditioned Natural Language Generation using only Unconditioned Language Model: An Exploration , 2020, ArXiv.
[6] François Laviolette,et al. Domain-Adversarial Training of Neural Networks , 2015, J. Mach. Learn. Res..
[7] Ziyan Wu,et al. Zero-Shot Deep Domain Adaptation , 2017, ECCV.
[8] John Blitzer,et al. Biographies, Bollywood, Boom-boxes and Blenders: Domain Adaptation for Sentiment Classification , 2007, ACL.
[9] Daniel Marcu,et al. Domain Adaptation for Statistical Classifiers , 2006, J. Artif. Intell. Res..
[10] Yi Yang,et al. Fast Easy Unsupervised Domain Adaptation with Marginalized Structured Dropout , 2014, ACL.
[11] Thomas Wolf,et al. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter , 2019, ArXiv.
[12] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[13] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[14] Roi Reichart,et al. Neural Structural Correspondence Learning for Domain Adaptation , 2016, CoNLL.
[15] Roi Reichart,et al. PERL: Pivot-based Domain Adaptation for Pre-trained Deep Contextualized Embedding Models , 2020, Transactions of the Association for Computational Linguistics.
[16] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[17] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[18] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[19] Sameer Singh,et al. Eliciting Knowledge from Language Models Using Automatically Generated Prompts , 2020, EMNLP.
[20] Hinrich Schütze,et al. FLORS: Fast and Simple Domain Adaptation for Part-of-Speech Tagging , 2014, TACL.
[21] Dean P. Foster,et al. TTI-TR-2009-1 Zero-Shot Domain Adaptation : A MultiView Approach , 2009 .
[22] Arkaitz Zubiaga,et al. Analysing How People Orient to and Spread Rumours in Social Media by Looking at Conversational Threads , 2015, PloS one.
[23] Shaogang Gong,et al. Unsupervised Domain Adaptation for Zero-Shot Learning , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[24] Alexander M. Rush,et al. Improved Parsing and POS Tagging Using Inter-Sentence Consistency Constraints , 2012, EMNLP-CoNLL.
[25] Claire Cardie,et al. Multinomial Adversarial Networks for Multi-Domain Text Classification , 2018, NAACL.
[26] Hal Daumé,et al. Frustratingly Easy Domain Adaptation , 2007, ACL.
[27] Isabelle Augenstein,et al. Transformer Based Multi-Source Domain Adaptation , 2020, EMNLP.
[28] Ashwin K. Vijayakumar,et al. Diverse Beam Search: Decoding Diverse Solutions from Neural Sequence Models , 2016, ArXiv.
[29] Brian Roark,et al. Supervised and unsupervised PCFG adaptation to novel domains , 2003, NAACL.
[30] Hui Xiong,et al. Transfer learning from multiple source domains via consensus regularization , 2008, CIKM '08.
[31] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[32] Arkaitz Zubiaga,et al. Exploiting Context for Rumour Detection in Social Media , 2017, SocInfo.
[33] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[34] Percy Liang,et al. Prefix-Tuning: Optimizing Continuous Prompts for Generation , 2021, ACL.
[35] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[36] Eugene Charniak,et al. Automatic Domain Adaptation for Parsing , 2010, NAACL.
[37] Jacob Eisenstein,et al. Unsupervised Domain Adaptation of Contextualized Embeddings for Sequence Labeling , 2019, EMNLP.
[38] Kilian Q. Weinberger,et al. Marginalized Denoising Autoencoders for Domain Adaptation , 2012, ICML.
[39] Roi Reichart,et al. Pivot Based Language Modeling for Improved Neural Domain Adaptation , 2018, NAACL.
[40] Lysandre Debut,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[41] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[42] Young-Bum Kim,et al. Domain Attention with an Ensemble of Experts , 2017, ACL.
[43] ChengXiang Zhai,et al. Instance Weighting for Domain Adaptation in NLP , 2007, ACL.
[44] Yoshua Bengio,et al. Domain Adaptation for Large-Scale Sentiment Classification: A Deep Learning Approach , 2011, ICML.
[45] Regina Barzilay,et al. Multi-Source Domain Adaptation with Mixture of Experts , 2018, EMNLP.
[46] Graham Neubig,et al. How Can We Know What Language Models Know? , 2019, Transactions of the Association for Computational Linguistics.
[47] José M. F. Moura,et al. Multiple Source Domain Adaptation with Adversarial Learning , 2018, ICLR.