A Latent Variable Model Approach to PMI-based Word Embeddings
暂无分享,去创建一个
Sanjeev Arora | Yuanzhi Li | Yingyu Liang | Tengyu Ma | Andrej Risteski | Sanjeev Arora | Yuanzhi Li | Tengyu Ma | Yingyu Liang | Andrej Risteski
[1] Geoffrey E. Hinton,et al. Three new graphical models for statistical language modelling , 2007, ICML '07.
[2] Elie Bienenstock,et al. Sphere Embedding: An Application to Part-of-Speech Induction , 2010, NIPS.
[3] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[4] James L. McClelland,et al. Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations , 1986 .
[5] Sham M. Kakade,et al. A Linear Dynamical System Model for Text , 2015, ICML.
[6] James L. McClelland,et al. James L. McClelland, David Rumelhart and the PDP Research Group, Parallel distributed processing: explorations in the microstructure of cognition . Vol. 1. Foundations . Vol. 2. Psychological and biological models . Cambridge MA: M.I.T. Press, 1987. , 1989, Journal of Child Language.
[7] Santosh S. Vempala,et al. Latent semantic indexing: a probabilistic analysis , 1998, PODS '98.
[8] Richard A. Harshman,et al. Indexing by Latent Semantic Analysis , 1990, J. Am. Soc. Inf. Sci..
[9] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[10] Tommi S. Jaakkola,et al. Word Embeddings as Metric Recovery in Semantic Spaces , 2016, TACL.
[11] Thomas Hofmann,et al. Probabilistic Latent Semantic Analysis , 1999, UAI.
[12] Yoram Singer,et al. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..
[13] Jason Weston,et al. A unified architecture for natural language processing: deep neural networks with multitask learning , 2008, ICML '08.
[14] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[15] Yoshua Bengio,et al. Neural Probabilistic Language Models , 2006 .
[16] John D. Lafferty,et al. Dynamic topic models , 2006, ICML.
[17] Dan Klein,et al. When and why are log-linear models self-normalizing? , 2015, NAACL.
[18] Douglas L. T. Rohde,et al. An Improved Model of Semantic Similarity Based on Lexical Co-Occurrence , 2005 .
[19] Omer Levy,et al. Linguistic Regularities in Sparse and Explicit Word Representations , 2014, CoNLL.
[20] Christopher Potts,et al. Learning Word Vectors for Sentiment Analysis , 2011, ACL.
[21] Patrick Pantel,et al. From Frequency to Meaning: Vector Space Models of Semantics , 2010, J. Artif. Intell. Res..
[22] Omer Levy,et al. Neural Word Embedding as Implicit Matrix Factorization , 2014, NIPS.
[23] Geoffrey Zweig,et al. Linguistic Regularities in Continuous Space Word Representations , 2013, NAACL.
[24] F. Black,et al. The Pricing of Options and Corporate Liabilities , 1973, Journal of Political Economy.
[25] Karl Stratos,et al. Spectral Learning of Latent-Variable PCFGs , 2012, ACL.
[26] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[27] J. Jenkins,et al. Word association norms , 1964 .
[28] Kenneth Ward Church,et al. Word Association Norms, Mutual Information, and Lexicography , 1989, ACL.
[29] Sanjeev Arora,et al. Linear Algebraic Structure of Word Senses, with Applications to Polysemy , 2016, TACL.
[30] David M. Blei,et al. Probabilistic topic models , 2012, Commun. ACM.
[31] Dean Alderucci. A SPECTRAL ALGORITHM FOR LEARNING HIDDEN MARKOV MODELS THAT HAVE SILENT STATES , 2015 .
[32] Gal Chechik,et al. Euclidean Embedding of Co-occurrence Data , 2004, J. Mach. Learn. Res..