BIDI: A classification algorithm with instance difficulty invariance

Abstract In artificial intelligence, an expert/intelligent systems can emulate the decision-making ability of human experts. A good classification algorithm can provide significant assistance to expert/intelligent systems in solving a variety of practical problems. In classification, the “hard” instances may be outliers or noisy instances that are difficult to learn, which may confuse the classifier and induce the overfitting problem in the case of placing much emphasis on them. In fact, the difficulty of instances is crucial for improving the generalization and credibility of classification. Unfortunately, nearly all the existing classifiers ignore this important information. In this paper, the classification difficulty of each instance is introduced from a statistical perspective, which is an inherent characteristic of the instance itself. Then, a new classification algorithm named “boosting with instance difficulty invariance (BIDI)” is proposed by incorporating the classification difficulty of instances. The BIDI conforms to the human cognition that easy instances are misclassified with a lower probability than difficult ones, and performs better with respect to generalization. The key insight of BIDI can provide relevant guidance for researchers to improve the generalization and credibility of classifiers in the expert systems of decision support systems. Experimental results demonstrate the effectiveness of BIDI in real-world data sets, indicating that it has great potential for solving many classification tasks of expert systems such as disease diagnosis and credit card fraud detection. Although the classification difficulty has strong statistical significance, its implementation remains computationally expensive. A fast method demonstrating rationality and feasibility is also proposed to approximate instances’ classification difficulty.

[1]  Ekrem Duman,et al.  A cost-sensitive decision tree approach for fraud detection , 2013, Expert Syst. Appl..

[2]  Hong Zhao,et al.  A cost sensitive decision tree algorithm based on weighted class distribution with batch deleting attribute mechanism , 2017, Inf. Sci..

[3]  Natarajan Sriraam,et al.  Classification of focal and non-focal EEG signals using neighborhood component analysis and machine learning algorithms , 2018, Expert Syst. Appl..

[4]  Juan José Rodríguez Diez,et al.  Diversity techniques improve the performance of the best imbalance learning ensembles , 2015, Inf. Sci..

[5]  Adolfo Martínez Usó,et al.  Making Sense of Item Response Theory in Machine Learning , 2016, ECAI.

[6]  C. Spearman The proof and measurement of association between two things. By C. Spearman, 1904. , 1987, The American journal of psychology.

[7]  Mehryar Mohri,et al.  Deep Boosting , 2014, ICML.

[8]  Zachary Chase Lipton A Critical Review of Recurrent Neural Networks for Sequence Learning , 2015, ArXiv.

[9]  Michael Granitzer,et al.  Sequence classification for credit-card fraud detection , 2018, Expert Syst. Appl..

[10]  Tony R. Martinez,et al.  An instance level analysis of data complexity , 2014, Machine Learning.

[11]  Yoav Freund,et al.  Boosting: Foundations and Algorithms , 2012 .

[12]  Tin Kam Ho,et al.  Complexity Measures of Supervised Classification Problems , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[13]  Pedro M. Domingos,et al.  On the Optimality of the Simple Bayesian Classifier under Zero-One Loss , 1997, Machine Learning.

[14]  Akhtar Hussain,et al.  An expert system for acoustic diagnosis of power circuit breakers and on-load tap changers , 2015, Expert Syst. Appl..

[15]  K. Pearson NOTES ON THE HISTORY OF CORRELATION , 1920 .

[16]  Yuncong Feng,et al.  A classification performance measure considering the degree of classification difficulty , 2016, Neurocomputing.

[17]  Julio Barbancho,et al.  Non-sequential automatic classification of anuran sounds for the estimation of climate-change indicators , 2018, Expert Syst. Appl..

[18]  Roberto Hornero,et al.  Utility of AdaBoost to Detect Sleep Apnea-Hypopnea Syndrome From Single-Channel Airflow , 2016, IEEE Transactions on Biomedical Engineering.

[19]  Jane You,et al.  Progressive subspace ensemble learning , 2016, Pattern Recognit..

[20]  Sreerama K. Murthy,et al.  Automatic Construction of Decision Trees from Data: A Multi-Disciplinary Survey , 1998, Data Mining and Knowledge Discovery.

[21]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1997, EuroCOLT.

[22]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[23]  T. Ho,et al.  Data Complexity in Pattern Recognition , 2006 .

[24]  Noel Lopes,et al.  Support Vector Machines (SVMs) , 2015 .

[25]  Haibin Ling,et al.  Exclusivity Regularized Machine: A New Ensemble SVM Classifier , 2017, IJCAI.

[26]  Kevin P. Murphy,et al.  Machine learning - a probabilistic perspective , 2012, Adaptive computation and machine learning series.

[27]  Vincent Lepetit,et al.  Learning Image Descriptors with Boosting , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[28]  David H. Wolpert,et al.  No free lunch theorems for optimization , 1997, IEEE Trans. Evol. Comput..

[29]  Gang Wang,et al.  An efficient diagnosis system for detection of Parkinson's disease using fuzzy k-nearest neighbor approach , 2013, Expert Syst. Appl..