Transfer Fine-Tuning: A BERT Case Study
暂无分享,去创建一个
[1] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[2] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[3] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[4] David Kauchak,et al. Improving Text Simplification Language Modeling Using Unsimplified Text Data , 2013, ACL.
[5] Christopher D. Manning,et al. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks , 2015, ACL.
[6] Luke S. Zettlemoyer,et al. Syntactic Scaffolds for Semantic Structures , 2018, EMNLP.
[7] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[8] Samuel R. Bowman,et al. Neural Network Acceptability Judgments , 2018, Transactions of the Association for Computational Linguistics.
[9] Sanja Fidler,et al. Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[10] Christopher Joseph Pal,et al. Learning General Purpose Distributed Sentence Representations via Large Scale Multi-task Learning , 2018, ICLR.
[11] Yulia Tsvetkov,et al. Style Transfer Through Back-Translation , 2018, ACL.
[12] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[13] Jun'ichi Tsujii,et al. Monolingual Phrase Alignment on Parse Forests , 2017, EMNLP.
[14] Chris Quirk,et al. Unsupervised Construction of Large Paraphrase Corpora: Exploiting Massively Parallel News Sources , 2004, COLING.
[15] Kevin Gimpel,et al. Pushing the Limits of Paraphrastic Sentence Embeddings with Millions of Machine Translations , 2017, ArXiv.
[16] Wei Wu,et al. Phrase-level Self-Attention Networks for Universal Sentence Encoding , 2018, EMNLP.
[17] Xiaodong Liu,et al. Multi-Task Deep Neural Networks for Natural Language Understanding , 2019, ACL.
[18] Eneko Agirre,et al. SemEval-2017 Task 1: Semantic Textual Similarity Multilingual and Crosslingual Focused Evaluation , 2017, *SEMEVAL.
[19] Jimmy J. Lin,et al. Pairwise Word Interaction Modeling with Deep Neural Networks for Semantic Similarity Measurement , 2016, NAACL.
[20] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[21] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[22] Samuel R. Bowman,et al. Sentence Encoders on STILTs: Supplementary Training on Intermediate Labeled-data Tasks , 2018, ArXiv.
[23] Hua He,et al. A Continuously Growing Dataset of Sentential Paraphrases , 2017, EMNLP.
[24] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[25] Holger Schwenk,et al. Supervised Learning of Universal Sentence Representations from Natural Language Inference Data , 2017, EMNLP.
[26] Myle Ott,et al. Understanding Back-Translation at Scale , 2018, EMNLP.
[27] Honglak Lee,et al. An efficient framework for learning sentence representations , 2018, ICLR.
[28] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[29] Zhen-Hua Ling,et al. Enhanced LSTM for Natural Language Inference , 2016, ACL.
[30] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[31] Sanja Fidler,et al. Skip-Thought Vectors , 2015, NIPS.
[32] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[33] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .