暂无分享,去创建一个
[1] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[2] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[3] Tal Linzen,et al. Issues in evaluating semantic spaces using word analogies , 2016, RepEval@ACL.
[4] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[5] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[6] Iryna Gurevych,et al. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks , 2019, EMNLP.
[7] Allyson Ettinger,et al. Assessing Composition in Sentence Vector Representations , 2018, COLING.
[8] Nan Hua,et al. Universal Sentence Encoder , 2018, ArXiv.
[9] Sanja Fidler,et al. Skip-Thought Vectors , 2015, NIPS.
[10] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[11] Matteo Pagliardini,et al. Unsupervised Learning of Sentence Embeddings Using Compositional n-Gram Features , 2017, NAACL.
[12] Samuel R. Bowman,et al. The Multi-Genre NLI Corpus , 2018 .
[13] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[14] Christopher Joseph Pal,et al. Learning General Purpose Distributed Sentence Representations via Large Scale Multi-task Learning , 2018, ICLR.
[15] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[16] Honglak Lee,et al. An efficient framework for learning sentence representations , 2018, ICLR.
[17] Yonatan Belinkov,et al. Fine-grained Analysis of Sentence Embeddings Using Auxiliary Prediction Tasks , 2016, ICLR.
[18] Christopher Potts,et al. A large annotated corpus for learning natural language inference , 2015, EMNLP.
[19] Yiming Yang,et al. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context , 2019, ACL.
[20] Michael L. Littman,et al. Corpus-based Learning of Analogies and Semantic Relations , 2005, Machine Learning.
[21] Anna Rumshisky,et al. A Primer in BERTology: What We Know About How BERT Works , 2020, Transactions of the Association for Computational Linguistics.
[22] Samuel R. Bowman,et al. Verb Argument Structure Alternations in Word and Sentence Embeddings , 2018, ArXiv.
[23] Yonatan Belinkov,et al. Proceedings of the 2018 EMNLP Workshop BlackboxNLP : Analyzing and Interpreting Neural Networks for NLP , 2018 .
[24] Gerard de Melo,et al. Exploring Semantic Properties of Sentence Embeddings , 2018, ACL.
[25] Geoffrey Zweig,et al. Linguistic Regularities in Continuous Space Word Representations , 2013, NAACL.
[26] Holger Schwenk,et al. Supervised Learning of Universal Sentence Representations from Natural Language Inference Data , 2017, EMNLP.
[27] Mona T. Diab,et al. Efficient Sentence Embedding using Discrete Cosine Transform , 2019, EMNLP/IJCNLP.
[28] Noah D. Goodman,et al. Evaluating Compositionality in Sentence Embeddings , 2018, CogSci.
[29] M. Marelli,et al. SemEval-2014 Task 1: Evaluation of Compositional Distributional Semantic Models on Full Sentences through Semantic Relatedness and Textual Entailment , 2014, *SEMEVAL.
[30] Natalie Schluter,et al. The Word Analogy Testing Caveat , 2018, NAACL.
[31] Omer Levy,et al. Linguistic Regularities in Sparse and Explicit Word Representations , 2014, CoNLL.
[32] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[33] Guillaume Lample,et al. What you can cram into a single $&!#* vector: Probing sentence embeddings for linguistic properties , 2018, ACL.
[34] Iryna Gurevych,et al. Concatenated Power Mean Word Embeddings as Universal Cross-Lingual Sentence Representations , 2018, 1803.01400.
[35] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[36] Malvina Nissim,et al. Fair is Better than Sensational: Man is to Doctor as Woman is to Doctor , 2019, Computational Linguistics.
[37] Terry Regier,et al. Probing sentence embeddings for structure-dependent tense , 2018, BlackboxNLP@EMNLP.