Efficient L1 Regularized Logistic Regression

L1 regularized logistic regression is now a workhorse of machine learning: it is widely used for many classification problems, particularly ones with many features. L1 regularized logistic regression requires solving a convex optimization problem. However, standard algorithms for solving convex optimization problems do not scale well enough to handle the large datasets encountered in many practical settings. In this paper, we propose an efficient algorithm for L1 regularized logistic regression. Our algorithm iteratively approximates the objective function by a quadratic approximation at the current point, while maintaining the L1 constraint. In each iteration, it uses the efficient LARS (Least Angle Regression) algorithm to solve the resulting L1 constrained quadratic optimization problem. Our theoretical results show that our algorithm is guaranteed to converge to the global optimum. Our experiments show that our algorithm significantly outperforms standard algorithms for solving convex optimization problems. Moreover, our algorithm outperforms four previously published algorithms that were specifically designed to solve the L1 regularized logistic regression problem.

[1]  丸山 徹 Convex Analysisの二,三の進展について , 1977 .

[2]  P. Green Iteratively reweighted least squares for maximum likelihood estimation , 1984 .

[3]  P. McCullagh,et al.  Generalized Linear Models , 1992 .

[4]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[5]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[6]  M. R. Osborne,et al.  On the LASSO and its Dual , 2000 .

[7]  Eric R. Ziegel,et al.  Generalized Linear Models , 2002, Technometrics.

[8]  James Theiler,et al.  Online Feature Selection using Grafting , 2003, ICML.

[9]  R. Tibshirani,et al.  Least angle regression , 2004, math/0406456.

[10]  Joshua Goodman,et al.  Exponential Priors for Maximum Entropy Models , 2004, NAACL.

[11]  Volker Roth,et al.  The generalized LASSO , 2004, IEEE Transactions on Neural Networks.

[12]  T. Minka A comparison of numerical optimizers for logistic regression , 2004 .

[13]  A. Ng Feature selection, L1 vs. L2 regularization, and rotational invariance , 2004, Twenty-first international conference on Machine learning - ICML '04.

[14]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.