Random Projection, Margins, Kernels, and Feature-Selection
暂无分享,去创建一个
[1] H. D. Block. The perceptron: a model for brain functioning. I , 1962 .
[2] Albert B Novikoff,et al. ON CONVERGENCE PROOFS FOR PERCEPTRONS , 1963 .
[3] Marvin Minsky,et al. Perceptrons: An Introduction to Computational Geometry , 1969 .
[4] Adi Ben-Israel,et al. Generalized inverses: theory and applications , 1974 .
[5] J. Meyer. Generalized Inverses (Theory And Applications) (Adi Ben-Israel and Thomas N. E. Greville) , 1976 .
[6] W. B. Johnson,et al. Extensions of Lipschitz mappings into Hilbert space , 1984 .
[7] Nick Littlestone,et al. From on-line to batch learning , 1989, COLT '89.
[8] Bernhard E. Boser,et al. A training algorithm for optimal margin classifiers , 1992, COLT '92.
[9] Yoav Freund,et al. A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.
[10] David P. Williamson,et al. Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming , 1995, JACM.
[11] Yoav Freund,et al. A decision-theoretic generalization of on-line learning and an application to boosting , 1997, EuroCOLT.
[12] Piotr Indyk,et al. Approximate nearest neighbors: towards removing the curse of dimensionality , 1998, STOC '98.
[13] John Shawe-Taylor,et al. Structural Risk Minimization Over Data-Dependent Hierarchies , 1998, IEEE Trans. Inf. Theory.
[14] Yoav Freund,et al. Large Margin Classification Using the Perceptron Algorithm , 1998, COLT' 98.
[15] Vladimir Vapnik,et al. Statistical learning theory , 1998 .
[16] John Shawe-Taylor,et al. Generalization Performance of Support Vector Machines and Other Pattern Classifiers , 1999 .
[17] Rafail Ostrovsky,et al. Efficient search for approximate nearest neighbor in high dimensional spaces , 1998, STOC '98.
[18] Santosh S. Vempala. Random projection: a new approach to VLSI layout , 1998, Proceedings 39th Annual Symposium on Foundations of Computer Science (Cat. No.98CB36280).
[19] Anupam Gupta,et al. An elementary proof of the Johnson-Lindenstrauss Lemma , 1999 .
[20] B. Schölkopf,et al. Advances in kernel methods: support vector learning , 1999 .
[21] Leonard J. Schulman,et al. Clustering for Edge-Cost Minimization , 1999, Electron. Colloquium Comput. Complex..
[22] Leonard J. Schulman,et al. Clustering for edge-cost minimization (extended abstract) , 2000, STOC '00.
[23] Sanjoy Dasgupta,et al. Experiments with Random Projection , 2000, UAI.
[24] Gunnar Rätsch,et al. An introduction to kernel-based learning algorithms , 2001, IEEE Trans. Neural Networks.
[25] Dimitris Achlioptas,et al. Database-friendly random projections , 2001, PODS.
[26] Dmitriy Fradkin,et al. Experiments with random projections for machine learning , 2003, KDD '03.
[27] Dimitris Achlioptas,et al. Database-friendly random projections: Johnson-Lindenstrauss with binary coins , 2003, J. Comput. Syst. Sci..
[28] Sanjoy Dasgupta,et al. An elementary proof of a theorem of Johnson and Lindenstrauss , 2003, Random Struct. Algorithms.
[29] Corinna Cortes,et al. Support-Vector Networks , 1995, Machine Learning.
[30] Santosh S. Vempala,et al. The Random Projection Method , 2005, DIMACS Series in Discrete Mathematics and Theoretical Computer Science.
[31] Santosh S. Vempala,et al. On Kernels, Margins, and Low-Dimensional Mappings , 2004, ALT.
[32] R. Schapire. The Strength of Weak Learnability , 1990, Machine Learning.
[33] Maria-Florina Balcan,et al. A PAC-Style Model for Learning from Labeled and Unlabeled Data , 2005, COLT.
[34] George Bebis,et al. Face recognition experiments with random projection , 2005, SPIE Defense + Commercial Sensing.
[35] Santosh S. Vempala,et al. An algorithmic theory of learning: Robust concepts and random projection , 1999, Machine Learning.
[36] Maria-Florina Balcan,et al. On a theory of learning with similarity functions , 2006, ICML.
[37] Santosh S. Vempala,et al. Kernels as features: On kernels, margins, and low-dimensional mappings , 2006, Machine Learning.