Multi-Kernel Multi-Label Learning with Max-Margin Concept Network

In this paper, a novel method is developed for enabling Multi-Kernel Multi-Label Learning. Interlabel dependency and similarity diversity are simultaneously leveraged in the proposed method. A concept network is constructed to capture the inter-label correlations for classifier training. Maximal margin approach is used to effectively formulate the feature-label associations and the labellabel correlations. Specific kernels are learned not only for each label but also for each pair of the inter-related labels. By learning the eigenfunctions of the kernels, the similarity between a new data point and the training samples can be computed in the online mode. Our experimental results on real datasets (web pages, images, music, and bioinformatics) have demonstrated the effectiveness of our method.

[1]  Thomas Hofmann,et al.  Support vector machine learning for interdependent and structured output spaces , 2004, ICML.

[2]  Zhi-Hua Zhou,et al.  ML-KNN: A lazy learning approach to multi-label learning , 2007, Pattern Recognit..

[3]  Tommi S. Jaakkola,et al.  More data means less inference: A pseudo-max approach to structured learning , 2010, NIPS.

[4]  Shi Bing,et al.  Inductive learning algorithms and representations for text categorization , 2006 .

[5]  Susan T. Dumais,et al.  Inductive learning algorithms and representations for text categorization , 1998, CIKM '98.

[6]  Leszek Wojnar,et al.  Image Analysis , 1998 .

[7]  Ben Taskar,et al.  Max-Margin Markov Networks , 2003, NIPS.

[8]  Jieping Ye,et al.  Learning the Optimal Neighborhood Kernel for Classification , 2009, IJCAI.

[9]  Andrew McCallum,et al.  Collective multi-label classification , 2005, CIKM '05.

[10]  Lihi Zelnik-Manor,et al.  Large Scale Max-Margin Multi-Label Classification with Priors , 2010, ICML.

[11]  Zhi-Hua Zhou,et al.  Multilabel dimensionality reduction via dependence maximization , 2008, TKDD.

[12]  Gerhard Winkler,et al.  Image analysis, random fields and dynamic Monte Carlo methods: a mathematical introduction , 1995, Applications of mathematics.

[13]  Grigorios Tsoumakas,et al.  Multi-Label Classification: An Overview , 2007, Int. J. Data Warehous. Min..

[14]  Tao Mei,et al.  Correlative multi-label video annotation , 2007, ACM Multimedia.

[15]  Jianping Fan,et al.  Structured max-margin learning for multi-label image annotation , 2010, CIVR '10.

[16]  T. Stephenson Image analysis , 1992, Nature.

[17]  Tony Jebara,et al.  Multi-task feature and kernel selection for SVMs , 2004, ICML.

[18]  Michael I. Jordan,et al.  Multiple kernel learning, conic duality, and the SMO algorithm , 2004, ICML.

[19]  Jiebo Luo,et al.  Learning multi-label scene classification , 2004, Pattern Recognit..

[20]  Chong-Wah Ngo,et al.  Semantic context modeling with maximal margin Conditional Random Fields for automatic image annotation , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[21]  Yves Grandvalet,et al.  More efficiency in multiple kernel learning , 2007, ICML '07.

[22]  Zhi-Hua Zhou,et al.  Multi-Label Learning with Weak Label , 2010, AAAI.

[23]  Jason Weston,et al.  A kernel method for multi-labelled classification , 2001, NIPS.

[24]  Rong Yan,et al.  Model-shared subspace boosting for multi-label classification , 2007, KDD '07.

[25]  Tibério S. Caetano,et al.  Reverse Multi-Label Learning , 2010, NIPS.

[26]  Gunnar Rätsch,et al.  Large Scale Multiple Kernel Learning , 2006, J. Mach. Learn. Res..