Fast SDP Relaxations of Graph Cut Clustering, Transduction, and Other Combinatorial Problem

The rise of convex programming has changed the face of many research fields in recent years, machine learning being one of the ones that benefitted the most. A very recent developement, the relaxation of combinatorial problems to semi-definite programs (SDP), has gained considerable attention over the last decade (Helmberg, 2000; De Bie and Cristianini, 2004a). Although SDP problems can be solved in polynomial time, for many relaxations the exponent in the polynomial complexity bounds is too high for scaling to large problem sizes. This has hampered their uptake as a powerful new tool in machine learning. In this paper, we present a new and fast SDP relaxation of the normalized graph cut problem, and investigate its usefulness in unsupervised and semi-supervised learning. In particular, this provides a convex algorithm for transduction, as well as approaches to clustering. We further propose a whole cascade of fast relaxations that all hold the middle between older spectral relaxations and the new SDP relaxation, allowing one to trade off computational cost versus relaxation accuracy. Finally, we discuss how the methodology developed in this paper can be applied to other combinatorial problems in machine learning, and we treat the max-cut problem as an example.

[1]  David P. Williamson,et al.  Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming , 1995, JACM.

[2]  Stephen P. Boyd,et al.  Semidefinite Programming , 1996, SIAM Rev..

[3]  Jitendra Malik,et al.  Normalized cuts and image segmentation , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[4]  Jos F. Sturm,et al.  A Matlab toolbox for optimization over symmetric cones , 1999 .

[5]  C. Helmberg Semidefinite Programming for Combinatorial Optimization , 2000 .

[6]  Franz Rendl,et al.  A Spectral Bundle Method for Semidefinite Programming , 1999, SIAM J. Optim..

[7]  Avrim Blum,et al.  Learning from Labeled and Unlabeled Data using Graph Mincuts , 2001, ICML.

[8]  Michael I. Jordan,et al.  On Spectral Clustering: Analysis and an algorithm , 2001, NIPS.

[9]  Nello Cristianini,et al.  Spectral Kernel Methods for Clustering , 2001, NIPS.

[10]  Bernhard Schölkopf,et al.  Cluster Kernels for Semi-Supervised Learning , 2002, NIPS.

[11]  Henry Wolkowicz,et al.  Strengthened semidefinite relaxations via a second lifting for the Max-Cut problem , 2002, Discret. Appl. Math..

[12]  Renato D. C. Monteiro,et al.  A nonlinear programming algorithm for solving semidefinite programs via low-rank factorization , 2003, Math. Program..

[13]  Nello Cristianini,et al.  Convex Methods for Transduction , 2003, NIPS.

[14]  Tomer Hertz,et al.  Computing Gaussian Mixture Models with EM Using Equivalence Constraints , 2003, NIPS.

[15]  Thorsten Joachims,et al.  Transductive Learning via Spectral Graph Partitioning , 2003, ICML.

[16]  Dan Klein,et al.  Spectral Learning , 2003, IJCAI.

[17]  Kevin J. Lang Finding good nearly balanced cuts in power law graphs , 2004 .

[18]  Nello Cristianini,et al.  Kernel methods for exploratory data analysis: a demonstration on text data , 2004 .

[19]  Johan A. K. Suykens,et al.  Learning from General Label Constraints , 2004, SSPR/SPR.

[20]  Renato D. C. Monteiro,et al.  Digital Object Identifier (DOI) 10.1007/s10107-004-0564-1 , 2004 .

[21]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.