A maximum partial entropy-based method for multiple-instance concept learning

Multiple instance (MI) learning aims at identifying the underlying concept from collectively labeled data. A training sample consists of a set, known as a bag, of unlabelled instances. The bag as a whole is labeled positive if at least one instance in the bag is positive, or negative otherwise. Given such training samples, the goal is to learn a description of the common instance(s) among the positive bags, i.e., the underlying concept that is responsible for the positive label. In this work, we introduce a learning scheme based on the notion of partial entropy for MI concept learning. Partial entropy accentuates the intra-class information by focusing on the information reflected from the positive class in proportion to the total entropy, maximization of which is to equalize the likelihoods of intra-class outcomes among the positive class, essentially reflecting the intended concept. When coupled with a distance-based probabilistic model for MI learning, it is equivalent to seeking out a concept estimate that equalizes the intra-class distances while the distance to negative bags is restrained. It produces patterns that are similar to at least one instance from each of the positive bags while dissimilar from all instances in negative bags. The generated patterns from the optimization process correspond to prototypical concepts. Maximum partial entropy is conceptually simple and experimental results on different MI datasets demonstrate its effectiveness in learning an explicit representation of the concept and its competitive performance when applied to classification tasks.

[1]  Zhi-Hua Zhou,et al.  Multi-Instance Learning Based Web Mining , 2005, Applied Intelligence.

[2]  Giancarlo Ruffo,et al.  Learning single and multiple instance decision tree for computer security applications , 2000 .

[3]  Wang Fei,et al.  Structural Domain Based Multiple Instance Learning for Predicting Gram-Positive Bacterial Protein Subcellular Localization , 2009, 2009 International Joint Conference on Bioinformatics, Systems Biology and Intelligent Computing.

[4]  Rujie Liu,et al.  Multi-graph multi-instance learning for object-based image and video retrieval , 2012, ICMR '12.

[5]  Zhi-Hua Zhou,et al.  Multi-instance learning by treating instances as non-I.I.D. samples , 2008, ICML '09.

[6]  William M. Smith,et al.  A Study of Thinking , 1956 .

[7]  James R. Foulds,et al.  A review of multi-instance learning assumptions , 2010, The Knowledge Engineering Review.

[8]  William M. Smith,et al.  A Study of Thinking , 1956 .

[9]  Jun Wang,et al.  Solving the Multiple-Instance Problem: A Lazy Learning Approach , 2000, ICML.

[10]  Weixiong Zhang,et al.  Promoter prediction based on a multiple instance learning scheme , 2010, BCB '10.

[11]  Peter Auer,et al.  On Learning From Multi-Instance Examples: Empirical Evaluation of a Theoretical Approach , 1997, ICML.

[12]  Zhi-Hua Zhou,et al.  Adapting RBF Neural Networks to Multi-Instance Learning , 2006, Neural Processing Letters.

[13]  Zhiquan Qi,et al.  Online multiple instance boosting for object detection , 2011, Neurocomputing.

[14]  Xin Xu,et al.  Logistic Regression and Boosting for Labeled Bags of Instances , 2004, PAKDD.

[15]  Yann Chevaleyre,et al.  Solving Multiple-Instance and Multiple-Part Learning Problems with Decision Trees and Rule Sets. Application to the Mutagenesis Problem , 2001, Canadian Conference on AI.

[16]  Ming-Hsuan Yang,et al.  Robust Object Tracking with Online Multiple Instance Learning , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[17]  Tao Xu,et al.  Constructing Target Concept in Multiple Instance Learning Using Maximum Partial Entropy , 2012, MLDM.

[18]  Pietro Perona,et al.  Multiple Component Learning for Object Detection , 2008, ECCV.

[19]  Murat Dundar,et al.  Bayesian multiple instance learning: automatic feature selection and inductive transfer , 2008, ICML '08.

[20]  Qi Zhang,et al.  EM-DD: An Improved Multiple-Instance Learning Technique , 2001, NIPS.

[21]  Adam Tauman Kalai,et al.  A Note on Learning from Multiple-Instance Examples , 2004, Machine Learning.

[22]  Kaspar Riesen,et al.  IAM Graph Database Repository for Graph Based Pattern Recognition and Machine Learning , 2008, SSPR/SPR.

[23]  Tomás Lozano-Pérez,et al.  A Framework for Multiple-Instance Learning , 1997, NIPS.

[24]  Tao Xu,et al.  A multiple instance learning based framework for semantic image segmentation , 2010, Multimedia Tools and Applications.

[25]  Mark Craven,et al.  Supervised versus multiple instance learning: an empirical comparison , 2005, ICML.

[26]  Mykola Pechenizkiy,et al.  HyDR-MI: A hybrid algorithm to reduce dimensionality in multiple instance learning , 2013, Inf. Sci..

[27]  Silviu Guiaşu,et al.  Information theory with applications , 1977 .

[28]  Thomas G. Dietterich,et al.  Solving the Multiple Instance Problem with Axis-Parallel Rectangles , 1997, Artif. Intell..

[29]  Tao Xu,et al.  Adaptive Kernel Diverse Density Estimate for Multiple Instance Learning , 2011, MLDM.

[30]  Greg Mori,et al.  Multiple instance real boosting with aggregation functions , 2012, Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012).

[31]  Jaume Amores,et al.  Multiple instance classification: Review, taxonomy and comparative study , 2013, Artif. Intell..

[32]  Tao Xu,et al.  Future directions in Multiple Instance Learning , 2013 .

[33]  Thomas Hofmann,et al.  Support Vector Machines for Multiple-Instance Learning , 2002, NIPS.

[34]  Paul D. Gader,et al.  Random set framework for multiple instance learning , 2011, Inf. Sci..

[35]  Horst Bischof,et al.  MIForests: Multiple-Instance Learning with Randomized Trees , 2010, ECCV.

[36]  Hongbin Zha,et al.  Adaptive p-posterior mixture-model kernels for multiple instance learning , 2008, ICML '08.

[37]  Thomas Gärtner,et al.  Multi-Instance Kernels , 2002, ICML.