Data Compression Measures for Meta-Learning Systems

An important issue in predictive modeling is model selection. This process is time consuming and can be simplified with meta-learning. However, meta-learning systems need appropriate data descriptors for proper functioning. One of them are data compression measures which can be extracted out of the instance selection methods. When we only need to estimate the classification accuracy of the model, the compression obtained from instance selection is a good approximator, but when we need to estimate other performance measures such as the precision and sensitivity then the quality of the estimated performance drops. To overcome this issue we propose a new type of compression measure: the balanced compression which is sensitive to the class label distribution and shows high correlation with precision and sensitivity of the final classifiers. We also show that the application of the balanced compression as a meta-learning descriptor allows for precise assessment of the model performance, as proved by the presented experimental evaluation.

[1]  B. L. Welch The generalisation of student's problems when several different population variances are involved. , 1947, Biometrika.

[2]  Marcin Blachnik,et al.  A Hybrid System with Regression Trees in Steel-Making Process , 2011, HAIS.

[3]  Miroslaw Kordos Optimization of Evolutionary Instance Selection , 2017, ICAISC.

[4]  Quan Sun,et al.  Pairwise meta-rules for better meta-learning-based algorithm ranking , 2013, Machine Learning.

[5]  Marcin Blachnik,et al.  Information Selection and Data Compression RapidMiner Library , 2016 .

[6]  Michal Kozielski,et al.  A Meta-Learning Approach to Methane Concentration Value Prediction , 2015, BDAS.

[7]  Aaron Klein,et al.  Efficient and Robust Automated Machine Learning , 2015, NIPS.

[8]  João Mendes-Moreira,et al.  Towards Automatic Generation of Metafeatures , 2016, PAKDD.

[9]  Lars Kotthoff,et al.  Auto-WEKA 2.0: Automatic model selection and hyperparameter optimization in WEKA , 2017, J. Mach. Learn. Res..

[10]  Welch Bl THE GENERALIZATION OF ‘STUDENT'S’ PROBLEM WHEN SEVERAL DIFFERENT POPULATION VARLANCES ARE INVOLVED , 1947 .

[11]  Francisco Herrera,et al.  Prototype Selection for Nearest Neighbor Classification: Taxonomy and Empirical Study , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[12]  Andreas Dengel,et al.  Meta-learning for evolutionary parameter optimization of classifiers , 2012, Machine Learning.

[13]  Marcin Blachnik,et al.  Instance Selection in Logical Rule Extraction for Regression Problems , 2013, ICAISC.

[14]  Marcin Blachnik,et al.  On the Relation Between kNN Accuracy and Dataset Compression Level , 2016, ICAISC.

[15]  C. G. Hilborn,et al.  The Condensed Nearest Neighbor Rule , 1967 .

[16]  Dennis L. Wilson,et al.  Asymptotic Properties of Nearest Neighbor Rules Using Edited Data , 1972, IEEE Trans. Syst. Man Cybern..

[17]  Marcin Blachnik,et al.  Instance Selection for Classifier Performance Estimation in Meta Learning , 2017, Entropy.