The Gaussian hare and the Laplacian tortoise: computability of squared-error versus absolute-error estimators

Since the time of Gauss, it has been generally accepted that ?2-methods of combining observations by minimizing sums of squared errors have significant computational advantages over earlier ?1-methods based on minimization of absolute errors advocated by Boscovich, Laplace and others. However, ?1-methods are known to have significant robustness advantages over f2-methods in many applications, and related quantile regression methods provide a useful, complementary approach to classical least-squares estimation of statistical models. Combining recent advances in interior point methods for solving linear programs with a new statistical preprocessing approach for ?1-type problems, we obtain a 10to 100-fold improvement in computational speeds over current (simplex-based) ?1-algorithms in large problems, demonstrating that ?1-methods can be made competitive with f2-methods in terms of computational speed throughout the entire range of problem sizes. Formal complexity results suggest that ?1-regression can be made faster than least-squares regression for n sufficiently large and p modest.

[1]  F. Y. Edgeworth XXII. On a new method of reducing observations relating to several quantities , 1888 .

[2]  Abraham Charnes,et al.  Optimal Estimation of Executive Compensation by Linear Programming , 1955 .

[3]  H. M. Wagner Linear Programming Techniques for Regression Analysis , 1959 .

[4]  M. M. Siddiqui Distribution of quantiles in samples from a bivariate population , 1960 .

[5]  Anthony V. Fiacco,et al.  Nonlinear programming;: Sequential unconstrained minimization techniques , 1968 .

[6]  S. R. Searle Linear Models , 1971 .

[7]  Ian Barrodale,et al.  Algorithm 478: Solution of an Overdetermined System of Equations in the l1 Norm [F4] , 1974, Commun. ACM.

[8]  Ronald L. Rivest,et al.  Expected time bounds for selection , 1975, Commun. ACM.

[9]  H. Oja Descriptive Statistics for Multivariate Distributions , 1983 .

[10]  W. Steiger,et al.  Least Absolute Deviations: Theory, Applications and Algorithms , 1984 .

[11]  Stephen M. Stigler,et al.  Studies in the history of probability and statistics XL Boscovich, Simpson and a 1760 manuscript note on fitting a linear relation , 1984 .

[12]  Narendra Karmarkar,et al.  A new polynomial-time algorithm for linear programming , 1984, Comb..

[13]  Michael A. Saunders,et al.  On projected newton barrier methods for linear programming and an equivalence to Karmarkar’s projective method , 1986, Math. Program..

[14]  J. Powell,et al.  Censored regression quantiles , 1986 .

[15]  R. Koenker,et al.  Computing regression quantiles , 1987 .

[16]  Ed Anderson,et al.  LAPACK Users' Guide , 1995 .

[17]  James Renegar,et al.  A polynomial-time algorithm, based on Newton's method, for linear programming , 1988, Math. Program..

[18]  P. Hall,et al.  On the Distribution of a Studentized Quantile , 1988 .

[19]  Josef Stoer,et al.  On the complexity of following the central path of linear programs by linear extrapolation II , 1991, Math. Program..

[20]  Stephen Portnoy,et al.  Asymptotic Behavior of the Number of Regression Quantile Breakpoints , 1991, SIAM J. Sci. Comput..

[21]  Gary Chamberlain,et al.  QUANTILE REGRESSION, CENSORING, AND THE STRUCTURE OF WAGES , 1991 .

[22]  Margaret H. Wright,et al.  Interior methods for constrained optimization , 1992, Acta Numerica.

[23]  Clóvis C. Gonzaga,et al.  Path-Following Methods for Linear Programming , 1992, SIAM Rev..

[24]  Sanjay Mehrotra,et al.  On the Implementation of a Primal-Dual Interior Point Method , 1992, SIAM J. Optim..

[25]  Roy E. Marsten,et al.  On Implementing Mehrotra's Predictor-Corrector Interior-Point Method for Linear Programming , 1992, SIAM J. Optim..

[26]  Yin Zhang Primal-dual interior point approach for computingl1-solutions andl∞-solutions of overdetermined linear systems , 1993 .

[27]  Ron Shamir,et al.  Probabilistic Analysis in Linear Programming , 1993 .

[28]  Roger Koenker,et al.  Tests of linear hypotheses based on regression rank scores , 1993 .

[29]  Shinji Mizuno,et al.  On Adaptive-Step Primal-Dual Interior-Point Algorithms for Linear Programming , 1993, Math. Oper. Res..

[30]  Moshe Buchinsky CHANGES IN THE U.S. WAGE STRUCTURE 1963-1987: APPLICATION OF QUANTILE REGRESSION , 1994 .

[31]  R. Koenker,et al.  A Remark on Algorithm as 229: Computing Dual Regression Quantiles and Regression Rank Scores , 1994 .

[32]  Jianqing Fan,et al.  Local polynomial modelling and its applications , 1994 .

[33]  Linear Programming Techniques , 1994 .

[34]  R. Koenker Confidence Intervals for Regression Quantiles , 1994 .

[35]  B. Silverman,et al.  Nonparametric regression and generalized linear models , 1994 .

[36]  Pin T. Ng,et al.  Quantile smoothing splines , 1994 .

[37]  Moshe Buchinsky,et al.  Quantile regression, Box-Cox transformation model, and the U.S. wage structure, 1963–1987 , 1995 .

[38]  W. Manning,et al.  The demand for alcohol: the differential response to price. , 1995, Journal of health economics.

[39]  G. Stewart,et al.  Theory of the combination of observations least subject to error : part one, part two, supplement = Theoria combinationis observationum erroribus minimus obnoxiae : pars prior, pars posterior, supplementum , 1995 .

[40]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .