暂无分享,去创建一个
[1] Roberto Navigli,et al. Neural Sequence Learning Models for Word Sense Disambiguation , 2017, EMNLP.
[2] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[3] Nigel Collier,et al. De-Conflated Semantic Representations , 2016, EMNLP.
[4] George A. Miller,et al. WordNet: A Lexical Database for English , 1995, HLT.
[5] Andrew Gordon Wilson,et al. Multimodal Word Distributions , 2017, ACL.
[6] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[7] Chris Dyer,et al. Ontology-Aware Token Embeddings for Prepositional Phrase Attachment , 2017, ACL.
[8] Rico Sennrich,et al. Why Self-Attention? A Targeted Evaluation of Neural Machine Translation Architectures , 2018, EMNLP.
[9] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[10] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[11] Pascal Vincent,et al. Auto-Encoding Dictionary Definitions into Consistent Word Embeddings , 2018, EMNLP.
[12] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[13] Evangelos Kanoulas,et al. Improving Word Embedding Compositionality using Lexicographic Definitions , 2018, WWW.
[14] Anson Bastos,et al. Learning sentence embeddings using Recursive Networks , 2018, ArXiv.
[15] Stan Matwin,et al. One Single Deep Bidirectional LSTM Network for Word Sense Disambiguation of Text Data , 2018, Canadian Conference on AI.
[16] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[17] Tomas Mikolov,et al. Enriching Word Vectors with Subword Information , 2016, TACL.
[18] Graham Neubig,et al. When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation? , 2018, NAACL.
[19] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[20] José Camacho-Collados,et al. From Word to Sense Embeddings: A Survey on Vector Representations of Meaning , 2018, J. Artif. Intell. Res..
[21] Sanjeev Arora,et al. Linear Algebraic Structure of Word Senses, with Applications to Polysemy , 2016, TACL.
[22] Yoshua Bengio,et al. Learning to Understand Phrases by Embedding the Dictionary , 2015, TACL.
[23] Nigel Collier,et al. Mapping Text to Knowledge Graph Entities using Multi-Sense LSTMs , 2018, EMNLP.
[24] Michael Zock,et al. Word Lookup on the Basis of Associations : from an Idea to a Roadmap , 2004 .
[25] Nigel Collier,et al. Towards a Seamless Integration of Word Senses into Downstream NLP Applications , 2017, ACL.
[26] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[27] John Liu,et al. sense2vec - A Fast and Accurate Method for Word Sense Disambiguation In Neural Word Embeddings , 2015, ArXiv.
[28] Hinrich Schütze,et al. AutoExtend: Extending Word Embeddings to Embeddings for Synsets and Lexemes , 2015, ACL.
[29] Luke S. Zettlemoyer,et al. Dissecting Contextual Word Embeddings: Architecture and Representation , 2018, EMNLP.