Automatic Tuning of the RBF Kernel Parameter for Batch-Mode Active Learning Algorithms: A Scalable Framework

Batch-mode active learning algorithms can select a batch of valuable unlabeled samples to manually annotate for reducing the total cost of labeling every unlabeled sample. To facilitate selection of valuable unlabeled samples, many batch-mode active learning algorithms map samples to the reproducing kernel Hilbert space induced by a radial-basis function (RBF) kernel. Setting a proper value to the parameter for the RBF kernel is crucial for such batch-mode active learning algorithms. In this paper, for automatic tuning of the kernel parameter, a hypothesis-margin-based criterion function is proposed. Three frameworks are also developed to incorporate the function of automatic tuning of the kernel parameter with existing batch-model active learning algorithms. In the proposed frameworks, the kernel parameter can be tuned in a single stage or in multiple stages. Tuning the kernel parameter in a single stage aims for the kernel parameter to be suitable for selecting the specified number of unlabeled samples. When the kernel parameter is tuned in multiple stages, the incorporated active learning algorithm can be enforced to make coarse-to-fine evaluations of the importance of unlabeled samples. The proposed framework can also improve the scalability of existing batch-mode active learning algorithms satisfying a decomposition property. Experimental results on data sets comprising hundreds to hundreds of thousands of samples have shown the feasibility of the proposed framework.

[1]  Xiaofei He,et al.  Laplacian Regularized D-Optimal Design for Active Learning and Its Application to Image Retrieval , 2010, IEEE Transactions on Image Processing.

[2]  Chun Chen,et al.  Active Learning Based on Locally Linear Reconstruction , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[3]  Cheng Li,et al.  Active learning on manifolds , 2014, Neurocomputing.

[4]  Lorenzo Bruzzone,et al.  Batch-Mode Active-Learning Methods for the Interactive Classification of Remote Sensing Images , 2011, IEEE Transactions on Geoscience and Remote Sensing.

[5]  Yijun Sun,et al.  Iterative RELIEF for Feature Weighting: Algorithms, Theories, and Applications , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[6]  Sethuraman Panchanathan,et al.  Batch Mode Active Sampling Based on Marginal Probability Distribution Matching , 2013, ACM Trans. Knowl. Discov. Data.

[7]  Sebastián Ventura,et al.  Evolutionary Strategy to Perform Batch-Mode Active Learning on Multi-Label Data , 2018, ACM Trans. Intell. Syst. Technol..

[8]  Bernhard Schölkopf,et al.  A Kernel Two-Sample Test , 2012, J. Mach. Learn. Res..

[9]  Rong Jin,et al.  Active query selection for semi-supervised clustering , 2008, 2008 19th International Conference on Pattern Recognition.

[10]  Xiaoqi He,et al.  Combining clustering coefficient-based active learning and semi-supervised learning on networked data , 2010, 2010 IEEE International Conference on Intelligent Systems and Knowledge Engineering.

[11]  Wen Gao,et al.  Multiple kernel active learning for image classification , 2009, 2009 IEEE International Conference on Multimedia and Expo.

[12]  Sayan Mukherjee,et al.  Choosing Multiple Parameters for Support Vector Machines , 2002, Machine Learning.

[13]  Ling Huang,et al.  Fast approximate spectral clustering , 2009, KDD.

[14]  Yatong Zhou,et al.  Analysis of the Distance Between Two Classes for Tuning SVM Hyperparameters , 2010, IEEE Transactions on Neural Networks.

[15]  Edoardo Pasolli,et al.  Ensemble Multiple Kernel Active Learning For Classification of Multisource Remote Sensing Data , 2015, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing.

[16]  Nikolaos Papanikolopoulos,et al.  Scalable Active Learning for Multiclass Image Classification , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[17]  Jinbo Bi,et al.  Active learning via transductive experimental design , 2006, ICML.

[18]  G. Golub,et al.  Eigenvalue computation in the 20th century , 2000 .

[19]  Naftali Tishby,et al.  Margin based feature selection - theory and algorithms , 2004, ICML.

[20]  S. Sathiya Keerthi,et al.  Efficient tuning of SVM hyperparameters using radius/margin bound and iterative algorithms , 2002, IEEE Trans. Neural Networks.

[21]  David G. Lowe,et al.  Scalable Nearest Neighbor Algorithms for High Dimensional Data , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[22]  Rong Jin,et al.  Active Learning by Querying Informative and Representative Examples , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[23]  Burr Settles,et al.  Active Learning Literature Survey , 2009 .

[24]  Arindam Banerjee,et al.  Active Semi-Supervision for Pairwise Constrained Clustering , 2004, SDM.

[25]  Neil D. Lawrence,et al.  Optimising Kernel Parameters and Regularisation Coefficients for Non-linear Discriminant Analysis , 2006, J. Mach. Learn. Res..

[26]  Jieping Ye,et al.  A two-stage linear discriminant analysis via QR-decomposition , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[27]  Melba M. Crawford,et al.  Multiple kernel active learning for robust geo-spatial image analysis , 2013, 2013 IEEE International Geoscience and Remote Sensing Symposium - IGARSS.

[28]  Fei Wang,et al.  Semisupervised Metric Learning by Maximizing Constraint Margin , 2011, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[29]  Daphne Koller,et al.  Support Vector Machine Active Learning with Applications to Text Classification , 2000, J. Mach. Learn. Res..

[30]  Koby Crammer,et al.  Margin Analysis of the LVQ Algorithm , 2002, NIPS.

[31]  Rong Jin,et al.  Batch mode active learning and its application to medical image classification , 2006, ICML.

[32]  Yuhong Guo,et al.  Active Instance Sampling via Matrix Partition , 2010, NIPS.

[33]  Sethuraman Panchanathan,et al.  Active Batch Selection via Convex Relaxations with Guaranteed Solution Bounds , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[34]  Abdulrahman H. Altalhi,et al.  Statistical comparisons of active learning strategies over multiple datasets , 2018, Knowl. Based Syst..

[35]  Rong Jin,et al.  Semisupervised SVM batch mode active learning with applications to image retrieval , 2009, TOIS.

[36]  Lorenzo Bruzzone,et al.  A Batch-Mode Active Learning Technique Based on Multiple Uncertainty for SVM Classifier , 2012, IEEE Geoscience and Remote Sensing Letters.

[37]  Yao Hu,et al.  Active learning via neighborhood reconstruction , 2013, IJCAI 2013.

[38]  Jieping Ye,et al.  Querying discriminative and representative samples for batch mode active learning , 2013, KDD.

[39]  Chun Chen,et al.  Manifold optimal experimental design via dependence maximization for active learning , 2014, Neurocomputing.

[40]  Lyle H. Ungar,et al.  Machine Learning manuscript No. (will be inserted by the editor) Active Learning for Logistic Regression: , 2007 .

[41]  Sebastián Ventura,et al.  Effective active learning strategy for multi-label learning , 2018, Neurocomputing.

[42]  Deng Cai,et al.  Manifold Adaptive Experimental Design for Text Categorization , 2012, IEEE Transactions on Knowledge and Data Engineering.