Boosting for Vote Learning in High-Dimensional kNN Classification
暂无分享,去创建一个
[1] Dunja Mladenic,et al. A probabilistic approach to nearest-neighbor classification: naive hubness bayesian kNN , 2011, CIKM '11.
[2] Richard Bellman,et al. Adaptive Control Processes: A Guided Tour , 1961, The Mathematical Gazette.
[3] Nenad Tomašev. THE INFLUENCE OF WEIGHTING THE K-OCCURRENCES ON HUBNESS-AWARE CLASSIFICATION METHODS , 2011 .
[4] Charu C. Aggarwal,et al. On the Surprising Behavior of Distance Metrics in High Dimensional Spaces , 2001, ICDT.
[5] Michel Verleysen,et al. The Concentration of Fractional Distances , 2007, IEEE Transactions on Knowledge and Data Engineering.
[6] Irina Rish,et al. An empirical study of the naive Bayes classifier , 2001 .
[7] Lars Schmidt-Thieme,et al. Time-Series Classification Based on Individualised Error Prediction , 2010, 2010 13th IEEE International Conference on Computational Science and Engineering.
[8] Olatz Arbelaitz,et al. A new algorithm to build consolidated trees: study of the error rate and steadiness , 2004, Intelligent Information Systems.
[9] Laurent Amsaleg,et al. Locality sensitive hashing: A comparison of hash function types and querying mechanisms , 2010, Pattern Recognit. Lett..
[10] Oskar Söderkvist,et al. Computer Vision Classification of Leaves from Swedish Trees , 2001 .
[11] François Pachet,et al. Improving Timbre Similarity : How high’s the sky ? , 2004 .
[12] Jonathan Goldstein,et al. When Is ''Nearest Neighbor'' Meaningful? , 1999, ICDT.
[13] Alexandros Nanopoulos,et al. On the existence of obstinate results in vector space models , 2010, SIGIR.
[14] Li Wei,et al. Intelligent Icons: Integrating Lite-Weight Data Mining and Visualization into GUI Operating Systems , 2006, Sixth International Conference on Data Mining (ICDM'06).
[15] Nathalie Japkowicz,et al. Boosting Support Vector Machines for Imbalanced Data Sets , 2008, ISMIS.
[16] Markus Schedl,et al. Using Mutual Proximity to Improve Content-Based Audio Similarity , 2011, ISMIR.
[17] Stan Matwin,et al. Inner Ensembles: Using Ensemble Methods Inside the Learning Algorithm , 2013, ECML/PKDD.
[18] Krisztian Buza,et al. Fusion Methods for Time-Series Classification , 2011 .
[19] Christos Faloutsos,et al. On the 'Dimensionality Curse' and the 'Self-Similarity Blessing' , 2001, IEEE Trans. Knowl. Data Eng..
[20] Li Wei,et al. Fast time series classification using numerosity reduction , 2006, ICML.
[21] Yousef Saad,et al. Fast Approximate kNN Graph Construction for High Dimensional Data via Recursive Lanczos Bisection , 2009, J. Mach. Learn. Res..
[22] Dunja Mladenic,et al. Nearest neighbor voting in high dimensional data: Learning from past occurrences , 2012, Comput. Sci. Inf. Syst..
[23] Dunja Mladenic,et al. Hub Co-occurrence Modeling for Robust High-Dimensional kNN Classification , 2013, ECML/PKDD.
[24] Günther Eibl,et al. Multiclass Boosting for Weak Classifiers , 2005, J. Mach. Learn. Res..
[25] Nenad Tomašev,et al. Hubness-based fuzzy measures for high-dimensional k-nearest neighbor classification , 2014 .
[26] Nicolás García-Pedrajas,et al. Boosting k-nearest neighbor classifier by means of input space projection , 2009, Expert Syst. Appl..
[27] Alexandros Nanopoulos,et al. Hubs in Space: Popular Nearest Neighbors in High-Dimensional Data , 2010, J. Mach. Learn. Res..
[28] Michel Verleysen,et al. The Curse of Dimensionality in Data Mining and Time Series Prediction , 2005, IWANN.
[29] Alexandros Nanopoulos,et al. Nearest neighbors in high-dimensional data: the emergence and influence of hubs , 2009, ICML '09.
[30] Rocco A. Servedio,et al. Adaptive Martingale Boosting , 2008, NIPS.
[31] Jun Zheng,et al. Cost-sensitive boosting neural networks for software defect prediction , 2010, Expert Syst. Appl..
[32] Karl Aberer,et al. Distributed similarity search in high dimensions using locality sensitive hashing , 2009, EDBT '09.
[33] J. L. Hodges,et al. Discriminatory Analysis - Nonparametric Discrimination: Consistency Properties , 1989 .
[34] James M. Keller,et al. A fuzzy K-nearest neighbor algorithm , 1985, IEEE Transactions on Systems, Man, and Cybernetics.
[35] Vikram Pudi,et al. Class Based Weighted K-Nearest Neighbor over Imbalance Dataset , 2013, PAKDD.
[36] Nuno Vasconcelos,et al. Cost-Sensitive Boosting , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[37] Dunja Mladenic,et al. Class imbalance and the curse of minority hubs , 2013, Knowl. Based Syst..
[38] Qinghua Hu,et al. Dynamic time warping constraint learning for large margin nearest neighbor classification , 2011, Inf. Sci..
[39] Dunja Mladenic,et al. Hubness-Aware Shared Neighbor Distances for High-Dimensional k-Nearest Neighbor Classification , 2012, HAIS.
[40] Carlotta Domeniconi,et al. Nearest neighbor ensemble , 2004, ICPR 2004.
[41] Chunhua Shen,et al. A direct formulation for totally-corrective multi-class boosting , 2011, CVPR 2011.
[42] B. Malek,et al. Novel Shoulder-Surfing Resistant Haptic-based Graphical Password , 2006 .
[43] Giorgio Ausiello,et al. Proceedings of the International Conference on Database Theory , 1986, ICDT 1986.
[44] Yang Wang,et al. Cost-sensitive boosting for classification of imbalanced data , 2007, Pattern Recognit..
[45] Rong Jin,et al. Multi-Class Learning by Smoothed Boosting , 2007, Machine Learning.
[46] Yoav Freund,et al. A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.
[47] Adam Tauman Kalai,et al. Potential-Based Agnostic Boosting , 2009, NIPS.
[48] Fabian Mörchen,et al. Discovering Interpretable Muscle Activation Patterns with the Temporal Data Mining Method , 2004, PKDD.
[49] Arthur Flexer,et al. Limitations of interactive music recommendation based on audio content , 2010, Audio Mostly Conference.