Boosting for Vote Learning in High-Dimensional kNN Classification

Intrinsically high-dimensional data has recently been shown to exhibit substantial hubness in terms of skewne ss of the k-nearest neighbor occurrence frequency distribution. While some points arise as centers of influence and dominate most k-nearest neighbor sets, other points occur very rarely and barely affect the inferred models. Hubness has been shown to be highly detrimental to many learning tasks and several hubnessaware learning methods have recently been proposed. This paper extends the existing fuzzy neighbor occurrence model s in order to enable cost-sensitive learning. We evaluate the ex tended implementations within the context of multi-class boosting, which is used to learn the appropriate neighbor votes during the reweighting iterations. The proposed approach is evaluated o n a series of high-dimensional datasets from various domains. The results demonstrate promising improvements of the propose d approach over the baselines.

[1]  Dunja Mladenic,et al.  A probabilistic approach to nearest-neighbor classification: naive hubness bayesian kNN , 2011, CIKM '11.

[2]  Richard Bellman,et al.  Adaptive Control Processes: A Guided Tour , 1961, The Mathematical Gazette.

[3]  Nenad Tomašev THE INFLUENCE OF WEIGHTING THE K-OCCURRENCES ON HUBNESS-AWARE CLASSIFICATION METHODS , 2011 .

[4]  Charu C. Aggarwal,et al.  On the Surprising Behavior of Distance Metrics in High Dimensional Spaces , 2001, ICDT.

[5]  Michel Verleysen,et al.  The Concentration of Fractional Distances , 2007, IEEE Transactions on Knowledge and Data Engineering.

[6]  Irina Rish,et al.  An empirical study of the naive Bayes classifier , 2001 .

[7]  Lars Schmidt-Thieme,et al.  Time-Series Classification Based on Individualised Error Prediction , 2010, 2010 13th IEEE International Conference on Computational Science and Engineering.

[8]  Olatz Arbelaitz,et al.  A new algorithm to build consolidated trees: study of the error rate and steadiness , 2004, Intelligent Information Systems.

[9]  Laurent Amsaleg,et al.  Locality sensitive hashing: A comparison of hash function types and querying mechanisms , 2010, Pattern Recognit. Lett..

[10]  Oskar Söderkvist,et al.  Computer Vision Classification of Leaves from Swedish Trees , 2001 .

[11]  François Pachet,et al.  Improving Timbre Similarity : How high’s the sky ? , 2004 .

[12]  Jonathan Goldstein,et al.  When Is ''Nearest Neighbor'' Meaningful? , 1999, ICDT.

[13]  Alexandros Nanopoulos,et al.  On the existence of obstinate results in vector space models , 2010, SIGIR.

[14]  Li Wei,et al.  Intelligent Icons: Integrating Lite-Weight Data Mining and Visualization into GUI Operating Systems , 2006, Sixth International Conference on Data Mining (ICDM'06).

[15]  Nathalie Japkowicz,et al.  Boosting Support Vector Machines for Imbalanced Data Sets , 2008, ISMIS.

[16]  Markus Schedl,et al.  Using Mutual Proximity to Improve Content-Based Audio Similarity , 2011, ISMIR.

[17]  Stan Matwin,et al.  Inner Ensembles: Using Ensemble Methods Inside the Learning Algorithm , 2013, ECML/PKDD.

[18]  Krisztian Buza,et al.  Fusion Methods for Time-Series Classification , 2011 .

[19]  Christos Faloutsos,et al.  On the 'Dimensionality Curse' and the 'Self-Similarity Blessing' , 2001, IEEE Trans. Knowl. Data Eng..

[20]  Li Wei,et al.  Fast time series classification using numerosity reduction , 2006, ICML.

[21]  Yousef Saad,et al.  Fast Approximate kNN Graph Construction for High Dimensional Data via Recursive Lanczos Bisection , 2009, J. Mach. Learn. Res..

[22]  Dunja Mladenic,et al.  Nearest neighbor voting in high dimensional data: Learning from past occurrences , 2012, Comput. Sci. Inf. Syst..

[23]  Dunja Mladenic,et al.  Hub Co-occurrence Modeling for Robust High-Dimensional kNN Classification , 2013, ECML/PKDD.

[24]  Günther Eibl,et al.  Multiclass Boosting for Weak Classifiers , 2005, J. Mach. Learn. Res..

[25]  Nenad Tomašev,et al.  Hubness-based fuzzy measures for high-dimensional k-nearest neighbor classification , 2014 .

[26]  Nicolás García-Pedrajas,et al.  Boosting k-nearest neighbor classifier by means of input space projection , 2009, Expert Syst. Appl..

[27]  Alexandros Nanopoulos,et al.  Hubs in Space: Popular Nearest Neighbors in High-Dimensional Data , 2010, J. Mach. Learn. Res..

[28]  Michel Verleysen,et al.  The Curse of Dimensionality in Data Mining and Time Series Prediction , 2005, IWANN.

[29]  Alexandros Nanopoulos,et al.  Nearest neighbors in high-dimensional data: the emergence and influence of hubs , 2009, ICML '09.

[30]  Rocco A. Servedio,et al.  Adaptive Martingale Boosting , 2008, NIPS.

[31]  Jun Zheng,et al.  Cost-sensitive boosting neural networks for software defect prediction , 2010, Expert Syst. Appl..

[32]  Karl Aberer,et al.  Distributed similarity search in high dimensions using locality sensitive hashing , 2009, EDBT '09.

[33]  J. L. Hodges,et al.  Discriminatory Analysis - Nonparametric Discrimination: Consistency Properties , 1989 .

[34]  James M. Keller,et al.  A fuzzy K-nearest neighbor algorithm , 1985, IEEE Transactions on Systems, Man, and Cybernetics.

[35]  Vikram Pudi,et al.  Class Based Weighted K-Nearest Neighbor over Imbalance Dataset , 2013, PAKDD.

[36]  Nuno Vasconcelos,et al.  Cost-Sensitive Boosting , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[37]  Dunja Mladenic,et al.  Class imbalance and the curse of minority hubs , 2013, Knowl. Based Syst..

[38]  Qinghua Hu,et al.  Dynamic time warping constraint learning for large margin nearest neighbor classification , 2011, Inf. Sci..

[39]  Dunja Mladenic,et al.  Hubness-Aware Shared Neighbor Distances for High-Dimensional k-Nearest Neighbor Classification , 2012, HAIS.

[40]  Carlotta Domeniconi,et al.  Nearest neighbor ensemble , 2004, ICPR 2004.

[41]  Chunhua Shen,et al.  A direct formulation for totally-corrective multi-class boosting , 2011, CVPR 2011.

[42]  B. Malek,et al.  Novel Shoulder-Surfing Resistant Haptic-based Graphical Password , 2006 .

[43]  Giorgio Ausiello,et al.  Proceedings of the International Conference on Database Theory , 1986, ICDT 1986.

[44]  Yang Wang,et al.  Cost-sensitive boosting for classification of imbalanced data , 2007, Pattern Recognit..

[45]  Rong Jin,et al.  Multi-Class Learning by Smoothed Boosting , 2007, Machine Learning.

[46]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[47]  Adam Tauman Kalai,et al.  Potential-Based Agnostic Boosting , 2009, NIPS.

[48]  Fabian Mörchen,et al.  Discovering Interpretable Muscle Activation Patterns with the Temporal Data Mining Method , 2004, PKDD.

[49]  Arthur Flexer,et al.  Limitations of interactive music recommendation based on audio content , 2010, Audio Mostly Conference.