Combining Uncertainty Sampling methods for supporting the generation of meta-examples

Meta-Learning aims to automatically acquire knowledge relating features of learning problems to the performance of learning algorithms. Each training example in Meta-Learning (i.e. each meta-example) stores features of a learning problem plus the performance obtained by a set of algorithms when evaluated on the problem. Based on a set of meta-examples, a Meta-Learner will be used to predict algorithm performance for new problems. The generation of a good set of meta-examples can be a costly process, since for each problem it is necessary to perform an empirical evaluation of the algorithms. In a previous work, we proposed the Active Meta-Learning, in which Active Learning was used to reduce the set of meta-examples by selecting only the most relevant problems for meta-example generation. In the current work, we extend our previous research by combining different Uncertainty Sampling methods for Active Meta-Learning, considering that each individual method will provide useful information to select relevant problems. We also investigated the use of Outlier Detection to remove a priori those problems considered as outliers, aiming to improve the performance of the sampling methods. In our experiments, we observed a gain in Meta-Learning performance when the proposed combining method was compared to the individual active methods being combined and also when outliers were removed from the set of problems available for meta-example generation.

[1]  Joaquin Vanschoren,et al.  Meta-Learning Architectures: Collecting, Organizing and Exploiting Meta-Knowledge , 2011, Meta-Learning in Computational Intelligence.

[2]  Hilan Bensusan,et al.  Estimating the Predictive Accuracy of a Classifier , 2001, ECML.

[3]  Mark Craven,et al.  An Analysis of Active Learning Strategies for Sequence Labeling Tasks , 2008, EMNLP.

[4]  R. Jones,et al.  Active Learning with Feedback on Both Features and Instances , 2006 .

[5]  João Gama,et al.  On Data and Algorithms: Understanding Inductive Performance , 2004, Machine Learning.

[6]  Andrew McCallum,et al.  Toward Optimal Active Learning through Sampling Estimation of Error Reduction , 2001, ICML.

[7]  Dianhui Wang,et al.  Extreme learning machines: a survey , 2011, Int. J. Mach. Learn. Cybern..

[8]  Xi-Zhao Wang,et al.  Improving Generalization of Fuzzy IF--THEN Rules by Maximizing Fuzzy Entropy , 2009, IEEE Transactions on Fuzzy Systems.

[9]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.

[10]  Daphne Koller,et al.  Active Learning for Parameter Estimation in Bayesian Networks , 2000, NIPS.

[11]  David A. Cohn,et al.  Improving generalization with active learning , 1994, Machine Learning.

[12]  Christian Rudolf Köpf,et al.  Meta-learning: strategies, implementations, and evaluations for algorithm selection , 2004 .

[13]  Felix Naumann,et al.  Data fusion , 2009, CSUR.

[14]  Dan Roth,et al.  Margin-based active learning for structured predictions , 2010, Int. J. Mach. Learn. Cybern..

[15]  Ion Muslea,et al.  Active Learning with Multiple Views , 2009, Encyclopedia of Data Warehousing and Mining.

[16]  Ricardo Vilalta,et al.  Introduction to the Special Issue on Meta-Learning , 2004, Machine Learning.

[17]  Greg Schohn,et al.  Less is More: Active Learning with Support Vector Machines , 2000, ICML.

[18]  Dana Angluin,et al.  Queries and concept learning , 1988, Machine Learning.

[19]  Ricardo Vilalta,et al.  Using Meta-Learning to Support Data Mining , 2004, Int. J. Comput. Sci. Appl..

[20]  Ricardo Vilalta,et al.  Metalearning - Applications to Data Mining , 2008, Cognitive Technologies.

[21]  Xizhao Wang,et al.  Maximum Ambiguity-Based Sample Selection in Fuzzy Decision Tree Induction , 2012, IEEE Transactions on Knowledge and Data Engineering.

[22]  Ivan G. Costa,et al.  Mining Rules for the Automatic Selection Process of Clustering Methods Applied to Cancer Gene Expression Data , 2009, ICANN.

[23]  Vincent Corruble,et al.  Acquiring the Preferences of New Users in Recommender Systems : The Role of Item Controversy , 2022 .

[24]  Pat Langley,et al.  Selection of Relevant Features and Examples in Machine Learning , 1997, Artif. Intell..

[25]  Carlos Soares,et al.  Ranking Learning Algorithms: Using IBL and Meta-Learning on Accuracy and Time Results , 2003, Machine Learning.

[26]  Teresa Bernarda Ludermir,et al.  Combining Uncertainty Sampling Methods for Active Meta-Learning , 2009, 2009 Ninth International Conference on Intelligent Systems Design and Applications.

[27]  Teresa Bernarda Ludermir,et al.  Selective generation of training examples in active meta-learning , 2008, Int. J. Hybrid Intell. Syst..

[28]  Carlos Soares UCI++: Improved Support for Algorithm Selection Using Datasetoids , 2009, PAKDD.

[29]  Naoki Abe,et al.  Query Learning Strategies Using Boosting and Bagging , 1998, ICML.

[30]  Hema Raghavan,et al.  Active Learning with Feedback on Features and Instances , 2006, J. Mach. Learn. Res..

[31]  André Carlos Ponce de Leon Ferreira de Carvalho,et al.  Meta-learning approach to gene expression data classification , 2009, Int. J. Intell. Comput. Cybern..

[32]  Raymond J. Mooney,et al.  Diverse ensembles for active learning , 2004, ICML.

[33]  Raymond T. Ng,et al.  A Unified Notion of Outliers: Properties and Computation , 1997, KDD.

[34]  Kate Smith-Miles,et al.  Towards insightful algorithm selection for optimisation using meta-learning concepts , 2008, 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence).

[35]  Carlos Soares,et al.  A Meta-Learning Method to Select the Kernel Width in Support Vector Regression , 2004, Machine Learning.

[36]  Kate Smith-Miles,et al.  Cross-disciplinary perspectives on meta-learning for algorithm selection , 2009, CSUR.

[37]  Prasad Tadepalli,et al.  Active Learning with Committees for Text Categorization , 1997, AAAI/IAAI.

[38]  Teresa Bernarda Ludermir,et al.  Meta-learning approaches to selecting time series models , 2004, Neurocomputing.

[39]  Pavel Brazdil,et al.  Predicting relative performance of classifiers from samples , 2005, ICML '05.

[40]  H. Sebastian Seung,et al.  Query by committee , 1992, COLT '92.

[41]  Marcílio Carlos Pereira de Souto,et al.  Selecting Machine Learning Algorithms Using the Ranking Meta-Learning Approach , 2011, Meta-Learning in Computational Intelligence.

[42]  Aurora Trinidad Ramirez Pozo,et al.  Selecting software reliability models with a neural network meta classifier , 2008, 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence).

[43]  Daphne Koller,et al.  Support Vector Machine Active Learning with Applications to Text Classification , 2000, J. Mach. Learn. Res..

[44]  Dilek Z. Hakkani-Tür,et al.  Active learning: theory and applications to automatic speech recognition , 2005, IEEE Transactions on Speech and Audio Processing.

[45]  Michael Lindenbaum,et al.  Selective Sampling for Nearest Neighbor Classifiers , 1999, Machine Learning.

[46]  William A. Gale,et al.  A sequential algorithm for training text classifiers , 1994, SIGIR '94.

[47]  Saso Dzeroski,et al.  Combining Classifiers with Meta Decision Trees , 2003, Machine Learning.

[48]  Teresa Bernarda Ludermir,et al.  Active Generation of Training Examples in Meta-Regression , 2009, ICANN.

[49]  Stefan Wrobel,et al.  Active Hidden Markov Models for Information Extraction , 2001, IDA.

[50]  Francisco de A. T. de Carvalho,et al.  ActiveCP: A Method for Speeding up User Preferences Acquisition in Collaborative Filtering Systems , 2002, SBIA.

[51]  Lutz Prechelt,et al.  A Set of Neural Network Benchmark Problems and Benchmarking Rules , 1994 .

[52]  Norbert Jankowski,et al.  Meta-Learning in Computational Intelligence , 2013, Meta-Learning in Computational Intelligence.