Bagging based ensemble transfer learning

Nowadays, transfer learning is one of the main research areas in machine learning that is helpful for labeling the data with low cost. In this paper, we propose a novel bagging-based ensemble transfer learning (BETL). The BETL framework includes three operations: Initiate, Update, and Integrate. In the Initiate operation, we use bootstrap sampling to divide the source data into many subsets, and add the labeled data from the target domain into these subsets separately so that the source data and the target data arrive at a reasonable ratio, then we learn as many initial classifiers as the elements of an ensemble. In the Update operation, we utilize the initial classifiers and an updateable classifier to repeatedly label the data that hasn’t been labeled yet in the target domain, and then, add the newly labeled data into the target domain to renew the updateable classifier. In the Integrate operation, we integrate the updated classifiers from each iteration into a pool to predict the labels of the test data via the majority vote strategy. In order to demonstrate the effectiveness of our method in the classification process, we conduct experiments on UCI data set, real world data set, and text data set. The results show that our method can effectively label the unlabeled data in the target domain, which greatly enhances the performance of target domain.

[1]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[2]  Antonino Staiano,et al.  Machine learning and soft computing for ICT security: an overview of current trends , 2011, Journal of Ambient Intelligence and Humanized Computing.

[3]  Christian Esposito,et al.  Smart Cloud Storage Service Selection Based on Fuzzy Logic, Theory of Evidence and Game Theory , 2016, IEEE Transactions on Computers.

[4]  Shotaro Akaho,et al.  TrBagg: A Simple Transfer Learning Method and its Application to Personalization in Collaborative Tagging , 2009, 2009 Ninth IEEE International Conference on Data Mining.

[5]  Wenyin Gong,et al.  Differential Evolution With Ranking-Based Mutation Operators , 2013, IEEE Transactions on Cybernetics.

[6]  Hui Li,et al.  Enhanced Differential Evolution With Adaptive Strategies for Numerical Optimization , 2011, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[7]  Wei Fan,et al.  Actively Transfer Domain Knowledge , 2008, ECML/PKDD.

[8]  Daniel Marcu,et al.  Domain Adaptation for Statistical Classifiers , 2006, J. Artif. Intell. Res..

[9]  Qiang Yang,et al.  Boosting for transfer learning , 2007, ICML '07.

[10]  Hussein A. Abbass,et al.  The use of coevolution and the artificial immune system for ensemble learning , 2011, Soft Comput..

[11]  Ludmila I. Kuncheva,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2004 .

[12]  Qiang Yang,et al.  A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.

[13]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques, 3rd Edition , 1999 .

[14]  Koby Crammer,et al.  Learning from Multiple Sources , 2006, NIPS.

[15]  Nikunj C. Oza,et al.  Online Ensemble Learning , 2000, AAAI/IAAI.

[16]  Jiawei Han,et al.  Knowledge transfer via multiple model local structure mapping , 2008, KDD.

[17]  Leslie Pack Kaelbling,et al.  Efficient Bayesian Task-Level Transfer Learning , 2007, IJCAI.

[18]  Belhadri Messabih,et al.  Effect of simple ensemble methods on protein secondary structure prediction , 2015, Soft Comput..

[19]  Ian H. Witten,et al.  Chapter 1 – What's It All About? , 2011 .

[20]  Chien-Cheng Lee,et al.  An improved boosting algorithm and its application to facial emotion recognition , 2011, Journal of Ambient Intelligence and Humanized Computing.

[21]  Francesco Palmieri,et al.  Modeling security requirements for cloud‐based system development , 2015, Concurr. Comput. Pract. Exp..

[22]  Masashi Sugiyama,et al.  Mixture Regression for Covariate Shift , 2006, NIPS.

[23]  Wenyin Gong,et al.  DE/BBO: a hybrid differential evolution with biogeography-based optimization for global numerical optimization , 2010, Soft Comput..

[24]  Wei Liu,et al.  Extending Semi-supervised Learning Methods for Inductive Transfer Learning , 2009, 2009 Ninth IEEE International Conference on Data Mining.

[25]  Susan T. Dumais,et al.  The Combination of Text Classifiers Using Reliability Indicators , 2016, Information Retrieval.