Mittens: an Extension of GloVe for Learning Domain-Specialized Representations
暂无分享,去创建一个
[1] Pascal Vincent,et al. The Difficulty of Training Deep Architectures and the Effect of Unsupervised Pre-Training , 2009, AISTATS.
[2] Iryna Gurevych,et al. Reporting Score Distributions Makes a Difference: Performance Study of LSTM-networks for Sequence Tagging , 2017, EMNLP.
[3] Christopher Potts,et al. On the Effective Use of Pretraining for Natural Language Inference , 2017, ArXiv.
[4] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[5] Nigel Collier,et al. Improved Semantic Representation for Domain-Specific Entities , 2016, BioNLP@ACL.
[6] Christopher Potts,et al. Learning Word Vectors for Sentiment Analysis , 2011, ACL.
[7] Yoshua Bengio,et al. Why Does Unsupervised Pre-training Help Deep Learning? , 2010, AISTATS.
[8] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[9] Mark Dredze,et al. Improving Lexical Embeddings with Semantic Knowledge , 2014, ACL.
[10] Gaël Varoquaux,et al. The NumPy Array: A Structure for Efficient Numerical Computation , 2011, Computing in Science & Engineering.
[11] Tomas Mikolov,et al. Enriching Word Vectors with Subword Information , 2016, TACL.
[12] Kent A. Spackman,et al. SNOMED RT: a reference terminology for health care , 1997, AMIA.
[13] Yoram Singer,et al. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..
[14] David Vandyke,et al. Counter-fitting Word Vectors to Linguistic Constraints , 2016, NAACL.
[15] Leo Breiman,et al. Random Forests , 2001, Machine Learning.