Efficient Generalized Fused Lasso and its Application to the Diagnosis of Alzheimer's Disease

Generalized fused lasso (GFL) penalizes variables with L1 norms based both on the variables and their pairwise differences. GFL is useful when applied to data where prior information is expressed using a graph over the variables. However, the existing GFL algorithms incur high computational costs and they do not scale to highdimensional problems. In this study, we propose a fast and scalable algorithm for GFL. Based on the fact that fusion penalty is the Lovasz extension of a cut function, we show that the key building block of the optimization is equivalent to recursively solving parametric graph-cut problems. Thus, we use a parametric flow algorithm to solve GFL in an efficient manner. Runtime comparisons demonstrated a significant speed-up compared with the existing GFL algorithms. By exploiting the scalability of the proposed algorithm, we formulated the diagnosis of Alzheimer's disease as GFL. Our experimental evaluations demonstrated that the diagnosis performance was promising and that the selected critical voxels were well structured i.e., connected, consistent according to cross-validation and in agreement with prior clinical knowledge.

[1]  Peter A. Bandettini,et al.  Does feature selection improve classification accuracy? Impact of sample size and feature selection on classification using anatomical magnetic resonance images , 2012, NeuroImage.

[2]  Vladimir Kolmogorov,et al.  What energy functions can be minimized via graph cuts? , 2002, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[3]  R. Tibshirani,et al.  The solution path of the generalized lasso , 2010, 1005.1971.

[4]  O. Nelles,et al.  An Introduction to Optimization , 1996, IEEE Antennas and Propagation Magazine.

[5]  I. Johnstone,et al.  Adapting to Unknown Smoothness via Wavelet Shrinkage , 1995 .

[6]  Robert E. Tarjan,et al.  A Fast Parametric Maximum Flow Algorithm and Applications , 1989, SIAM J. Comput..

[7]  Yong He,et al.  Discriminative analysis of early Alzheimer's disease using multi-modal imaging and multi-level characterization with multi-classifier (M3) , 2012, NeuroImage.

[8]  Jiayu Zhou,et al.  A multi-task learning formulation for predicting disease progression , 2011, KDD.

[9]  Francis R. Bach,et al.  Structured sparsity-inducing norms through submodular functions , 2010, NIPS.

[10]  Julien Mairal,et al.  Convex and Network Flow Optimization for Structured Sparsity , 2011, J. Mach. Learn. Res..

[11]  John Ashburner,et al.  A fast diffeomorphic image registration algorithm , 2007, NeuroImage.

[12]  S. Fujishige,et al.  The Minimum-Norm-Point Algorithm Applied to Submodular Function Minimization and Linear Programming , 2006 .

[13]  Jieping Ye,et al.  An efficient algorithm for a class of fused lasso problems , 2010, KDD.

[14]  Yoshinobu Kawahara,et al.  Efficient network-guided multi-locus association mapping with graph cuts , 2012, Bioinform..

[15]  Michael A. Saunders,et al.  USER’S GUIDE FOR SNOPT 5.3: A FORTRAN PACKAGE FOR LARGE-SCALE NONLINEAR PROGRAMMING , 2002 .

[16]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[17]  James B. Orlin,et al.  A Faster Strongly Polynomial Time Algorithm for Submodular Function Minimization , 2007, IPCO.

[18]  Stephen P. Boyd,et al.  Disciplined Convex Programming , 2006 .

[19]  Kazuyuki Aihara,et al.  Equivalence of convex minimization problems over base polytopes , 2012 .

[20]  James B. Orlin,et al.  A faster strongly polynomial time algorithm for submodular function minimization , 2007, Math. Program..

[21]  Daoqiang Zhang,et al.  Domain Transfer Learning for MCI Conversion Prediction , 2012, MICCAI.

[22]  Junzhou Huang,et al.  Learning with structured sparsity , 2009, ICML '09.

[23]  Julien Mairal,et al.  Optimization with Sparsity-Inducing Penalties , 2011, Found. Trends Mach. Learn..

[24]  Marc Teboulle,et al.  A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems , 2009, SIAM J. Imaging Sci..

[25]  R. Tibshirani,et al.  Sparsity and smoothness via the fused lasso , 2005 .

[26]  Yoshinobu Kawahara,et al.  Structured Convex Optimization under Submodular Constraints , 2013, UAI.

[27]  Yurii Nesterov,et al.  Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.

[28]  John Darzentas,et al.  Problem Complexity and Method Efficiency in Optimization , 1983 .

[29]  Shuiwang Ji,et al.  SLEP: Sparse Learning with Efficient Projections , 2011 .

[30]  Wotao Yin,et al.  Parametric Maximum Flow Algorithms for Fast Total Variation Minimization , 2009, SIAM J. Sci. Comput..

[31]  Satoru Fujishige,et al.  Submodular functions and optimization , 1991 .

[32]  R. Tibshirani,et al.  PATHWISE COORDINATE OPTIMIZATION , 2007, 0708.1485.