暂无分享,去创建一个
Vít Novotný | Eniafe Festus Ayetiran | Michal Stefánik | Petr Sojka | Petr Sojka | Michal Štefánik | E. F. Ayetiran | Vít Novotný
[1] Derek Greene,et al. Practical solutions to the problem of diagonal dominance in kernel document clustering , 2006, ICML.
[2] Hideki Nakayama,et al. Compressing Word Embeddings via Deep Compositional Code Learning , 2017, ICLR.
[3] Maximilian Lam,et al. Word2Bits - Quantized Word Vectors , 2018, ArXiv.
[4] David D. Lewis,et al. Reuters-21578 Text Categorization Test Collection, Distribution 1.0 , 1997 .
[5] Grigori Sidorov,et al. Soft Similarity and Soft Cosine Measure: Similarity of Features in Vector Space Model , 2014, Computación y Sistemas.
[6] Graham Neubig,et al. When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation? , 2018, NAACL.
[7] Yue Yin,et al. Preliminary Exploration of Formula Embedding for Mathematical Information Retrieval: can mathematical formulae be embedded like a natural language? , 2017, ArXiv.
[8] Michael Werman,et al. A Linear Time Histogram Metric for Improved SIFT Matching , 2008, ECCV.
[9] Thorsten Joachims,et al. Text Categorization with Support Vector Machines: Learning with Many Relevant Features , 1998, ECML.
[10] Charles Elkan,et al. Latent semantic indexing (LSI) fails for TREC collections , 2011, SKDD.
[11] Matt J. Kusner,et al. From Word Embeddings To Document Distances , 2015, ICML.
[12] Guido Zuccon,et al. Integrating and Evaluating Neural Word Embeddings in Information Retrieval , 2015, ADCS.
[13] Ken Lang,et al. NewsWeeder: Learning to Filter Netnews , 1995, ICML.
[14] Delphine Charlet,et al. SimBow at SemEval-2017 Task 3: Soft-Cosine Semantic Similarity between Questions for Community Question Answering , 2017, *SEMEVAL.
[15] Risto Miikkulainen,et al. Test Data , 2019, Encyclopedia of Machine Learning and Data Mining.
[16] Matthijs Douze,et al. FastText.zip: Compressing text classification models , 2016, ArXiv.
[17] Chris Buckley,et al. OHSUMED: an interactive retrieval evaluation and new large test collection for research , 1994, SIGIR '94.
[18] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[19] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[20] Oren Kurland,et al. Query Expansion Using Word Embeddings , 2016, CIKM.
[21] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[22] Gábor Berend. ℓ1 Regularization of Word Embeddings for Multi-Word Expression Identification , 2018, Acta Cybern..
[23] Vít Novotný,et al. Implementation Notes for the Soft Cosine Measure , 2018, CIKM.
[24] Yixin Chen,et al. Compressing Neural Networks with the Hashing Trick , 2015, ICML.
[25] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[26] Ana Margarida de Jesus,et al. Improving Methods for Single-label Text Categorization , 2007 .
[27] Parul Parashar,et al. Neural Networks in Machine Learning , 2014 .
[28] Anton van den Hengel,et al. Image-Based Recommendations on Styles and Substitutes , 2015, SIGIR.
[29] Y. Benjamini,et al. Controlling the false discovery rate: a practical and powerful approach to multiple testing , 1995 .
[30] Richard A. Harshman,et al. Indexing by Latent Semantic Analysis , 1990, J. Am. Soc. Inf. Sci..
[31] Christopher D. Manning,et al. Compression of Neural Machine Translation Models via Pruning , 2016, CoNLL.
[32] Michael Werman,et al. Fast and robust Earth Mover's Distances , 2009, 2009 IEEE 12th International Conference on Computer Vision.
[33] Petr Sojka,et al. Software Framework for Topic Modelling with Large Corpora , 2010 .
[34] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[35] A. Agresti,et al. Approximate is Better than “Exact” for Interval Estimation of Binomial Proportions , 1998 .
[36] Kenji Sagae,et al. Combining Distributed Vector Representations for Words , 2015, VS@HLT-NAACL.
[37] Hod Lipson,et al. Re-embedding words , 2013, ACL.
[38] Xueqi Cheng,et al. Sparse Word Embeddings Using ℓ1 Regularized Online Learning , 2016, IJCAI.
[39] Yan Song,et al. Learning Word Representations with Regularization from Prior Knowledge , 2017, CoNLL.
[40] Douglas W. Oard,et al. Tangent-CFT: An Embedding Model for Mathematical Formulas , 2019, ICTIR.
[41] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[42] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[43] Gerard Salton,et al. Term-Weighting Approaches in Automatic Text Retrieval , 1988, Inf. Process. Manag..
[44] Pradeep Ravikumar,et al. Word Mover’s Embedding: From Word2Vec to Document Embedding , 2018, EMNLP.
[45] Wei Yang,et al. A Simple Regularization-based Algorithm for Learning Cross-Domain Word Embeddings , 2017, EMNLP.
[46] Stephen E. Robertson,et al. Understanding inverse document frequency: on theoretical arguments for IDF , 2004, J. Documentation.
[47] Zhi Jin,et al. A Comparative Study on Regularization Strategies for Embedding-based Neural Networks , 2015, EMNLP.