Inverse extreme learning machine for learning with label proportions

In large-scale learning problem, the scalability of learning algorithms is usually the key factor affecting the algorithm practical performance, which is determined by both the time complexity of the learning algorithms and the amount of supervision information (i.e., labeled data). Learning with label proportions (LLP) is a new kind of machine learning problem which has drawn much attention in recent years. Different from the well-known supervised learning, LLP can estimate a classifier from groups of weakly labeled data, where only the positive/negative class proportions of each group are known. Due to its weak requirements for the input data, LLP presents a variety of real-world applications in almost all the fields involving anonymous data, like computer vision, fraud detection and spam filtering. However, even through the required labeled data is of a very small amount, LLP still suffers from the long execution time a lot due to the high time complexity of the learning algorithm itself. In this paper, we propose a very fast learning method based on inversing output scaling process and extreme learning machine, namely Inverse Extreme Learning Machine (IELM), to address the above issues. IELM can speed up the training process by order of magnitudes for large datasets, while achieving highly competitive classification accuracy with the existing methods at the same time. Extensive experiments demonstrate the significant speedup of the proposed method. We also demonstrate the feasibility of IELM with a case study in real-world setting: modeling image attributes based on ImageNet Object Attributes dataset.

[1]  Jiashi Feng,et al.  Multi-class learning from class proportions , 2013, Neurocomputing.

[2]  Xiaojin Zhu,et al.  Introduction to Semi-Supervised Learning , 2009, Synthesis Lectures on Artificial Intelligence and Machine Learning.

[3]  Xia Liu,et al.  Is Extreme Learning Machine Feasible? A Theoretical Assessment (Part I) , 2015, IEEE Trans. Neural Networks Learn. Syst..

[4]  Korris Fu-Lai Chung,et al.  Support vector machine with manifold regularization and partially labeling privacy protection , 2015, Inf. Sci..

[5]  Richard Nock,et al.  (Almost) No Label No Cry , 2014, NIPS.

[6]  Richard A. Tapia,et al.  Practical Methods of Optimization, Volume 2: Constrained Optimization (R. Fletcher) , 1984 .

[7]  Iñaki Inza,et al.  Learning from Proportions of Positive and Unlabeled Examples , 2017, Int. J. Intell. Syst..

[8]  Klaus-Robert Müller,et al.  Learning from label proportions in brain-computer interfaces: Online unsupervised learning with guarantees , 2017, PloS one.

[9]  Tao Chen,et al.  Object-Based Visual Sentiment Concept Analysis and Application , 2014, ACM Multimedia.

[10]  Ming-Syan Chen,et al.  Video Event Detection by Inferring Temporal Instance Labels , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[11]  Faezeh Toutounian,et al.  A new method for computing Moore-Penrose inverse matrices , 2009 .

[12]  John Platt,et al.  Probabilistic Outputs for Support vector Machines and Comparisons to Regularized Likelihood Methods , 1999 .

[13]  Guang-Bin Huang,et al.  Extreme learning machine: a new learning scheme of feedforward neural networks , 2004, 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541).

[14]  Iñaki Inza,et al.  A Novel Weakly Supervised Problem: Learning from Positive-Unlabeled Proportions , 2015, CAEPIA.

[15]  David R. Musicant,et al.  Learning from Aggregate Views , 2006, 22nd International Conference on Data Engineering (ICDE'06).

[16]  Moisés Goldszmidt,et al.  Properties and Benefits of Calibrated Classifiers , 2004, PKDD.

[17]  Tao Chen,et al.  Modeling Attributes from Category-Attribute Proportions , 2014, ACM Multimedia.

[18]  Nando de Freitas,et al.  Learning about Individuals from Group Statistics , 2005, UAI.

[19]  Aron Culotta,et al.  Domain Adaptation for Learning from Label Proportions Using Self-Training , 2016, IJCAI.

[20]  Zhiquan Qi,et al.  Learning With Label Proportions via NPSVM , 2017, IEEE Transactions on Cybernetics.

[21]  Stefan R ping SVM Classifier Estimation from Group Probabilities , 2010, ICML 2010.

[22]  Razvan C. Bunescu,et al.  Multiple instance learning for sparse positive bags , 2007, ICML '07.

[23]  David R. Musicant,et al.  Supervised Learning by Training on Aggregate Outputs , 2007, Seventh IEEE International Conference on Data Mining (ICDM 2007).

[24]  Alexander J. Smola,et al.  Estimating Labels from Label Proportions , 2009, J. Mach. Learn. Res..

[25]  Katharina Morik,et al.  Learning from Label Proportions by Optimizing Cluster Model Selection , 2011, ECML/PKDD.

[26]  Marco Loog,et al.  On classification with bags, groups and sets , 2014, Pattern Recognit. Lett..

[27]  Frank Nielsen,et al.  Loss factorization, weakly supervised learning and label noise robustness , 2016, ICML.

[28]  Bo Wang,et al.  Learning with label proportions based on nonparallel support vector machines , 2017, Knowl. Based Syst..

[29]  Christoph H. Lampert,et al.  Attribute-Based Classification for Zero-Shot Visual Object Categorization , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[30]  James Allen Fill,et al.  The Moore-Penrose Generalized Inverse for Sums of Matrices , 1999, SIAM J. Matrix Anal. Appl..

[31]  Bernhard Schölkopf,et al.  Introduction to Semi-Supervised Learning , 2006, Semi-Supervised Learning.

[32]  Mahdi Pakdaman Naeini,et al.  Binary Classifier Calibration Using an Ensemble of Near Isotonic Regression Models , 2015, 2016 IEEE 16th International Conference on Data Mining (ICDM).

[33]  Iñaki Inza,et al.  Fitting the data from embryo implantation prediction: Learning from label proportions , 2018, Statistical methods in medical research.

[34]  Katharina Morik,et al.  Distributed Traffic Flow Prediction with Label Proportions: From in-Network towards High Performance Computation with MPI , 2015, MUD@ICML.

[35]  Thomas Hofmann,et al.  Support Vector Machines for Multiple-Instance Learning , 2002, NIPS.

[36]  Fei-Fei Li,et al.  Attribute Learning in Large-Scale Datasets , 2010, ECCV Workshops.

[37]  Jitendra Malik,et al.  Learning to detect natural image boundaries using local brightness, color, and texture cues , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[38]  Rich Caruana,et al.  Predicting good probabilities with supervised learning , 2005, ICML.

[39]  Bianca Zadrozny,et al.  Transforming classifier scores into accurate multiclass probability estimates , 2002, KDD.

[40]  Iñaki Inza,et al.  Weak supervision and other non-standard classification problems: A taxonomy , 2016, Pattern Recognit. Lett..

[41]  Dong Liu,et al.  $\propto$SVM for learning with label proportions , 2013, ICML 2013.

[42]  Shih-Fu Chang,et al.  On Learning with Label Proportions , 2014, ArXiv.

[43]  Chee Kheong Siew,et al.  Extreme learning machine: Theory and applications , 2006, Neurocomputing.