On Dimensional Linguistic Properties of the Word Embedding Space
暂无分享,去创建一个
Vaibhav Kumar | Florian Metze | Vikas Raunak | Vivek K. Gupta | Florian Metze | Vikas Raunak | Vivek Gupta | Vaibhav Kumar
[1] Sanjeev Arora,et al. A Latent Variable Model Approach to PMI-based Word Embeddings , 2015, TACL.
[2] Laure Thompson,et al. The strange geometry of skip-gram with negative sampling , 2017, EMNLP.
[3] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[4] Sanjeev Arora,et al. Linear Algebraic Structure of Word Senses, with Applications to Polysemy , 2016, TACL.
[5] Praneeth Netrapalli,et al. P-SIF: Document Embeddings Using Partition Averaging , 2020, AAAI.
[6] Zi Yin,et al. On the Dimensionality of Word Embedding , 2018, NeurIPS.
[7] David Vandyke,et al. Counter-fitting Word Vectors to Linguistic Constraints , 2016, NAACL.
[8] Sanjeev Arora,et al. A Simple but Tough-to-Beat Baseline for Sentence Embeddings , 2017, ICLR.
[9] Pramod Viswanath,et al. All-but-the-Top: Simple and Effective Postprocessing for Word Representations , 2017, ICLR.
[10] Graham Neubig,et al. When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation? , 2018, NAACL.
[11] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[12] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[13] Eneko Agirre,et al. SemEval-2012 Task 6: A Pilot on Semantic Textual Similarity , 2012, *SEMEVAL.
[14] Tomas Mikolov,et al. Enriching Word Vectors with Subword Information , 2016, TACL.
[15] Manaal Faruqui,et al. Community Evaluation and Exchange of Word Vectors at wordvectors.org , 2014, ACL.
[16] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[17] Timothy M. Hospedales,et al. Analogies Explained: Towards Understanding Word Embeddings , 2019, ICML.
[18] Omer Levy,et al. Neural Word Embedding as Implicit Matrix Factorization , 2014, NIPS.
[19] Jorge Cadima,et al. Principal component analysis: a review and recent developments , 2016, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences.
[20] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[21] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[22] Douwe Kiela,et al. SentEval: An Evaluation Toolkit for Universal Sentence Representations , 2018, LREC.
[23] Vikas Raunak,et al. Effective Dimensionality Reduction for Word Embeddings , 2017, RepL4NLP@ACL.
[24] Guillaume Lample,et al. What you can cram into a single $&!#* vector: Probing sentence embeddings for linguistic properties , 2018, ACL.