The Fast-τ Estimator for Regression

Yohai and Zamar's τ-estimators of regression have excellent statistical properties but are nevertheless rarely used in practice because of a lack of available software and the general impression that τ-estimators are difficult to approximate. We will show, however, that the computational difficulties of approximating τ-estimators are similar in nature to those of the more popular S-estimators. The main goal of this article is to compare an approximating algorithm for τ-estimators based on random resampling with some alternative heuristic search algorithms. We show that the former is not only simpler, but that when enhanced by local improvement steps it generally outperforms the consider edheuristic search algorithms, even when the seheuristic algorithms also incorporate local improvement steps. Additionally, we show that the random resampling algorithm for approximating τ-estimators has favorable statistical properties compared to the analogous and widely used algorithms for S- and least trimmed squares estimators.

[1]  David J. Olive,et al.  Inconsistency of Resampling Algorithms for High-Breakdown Regression Estimators and a New Algorithm , 2002 .

[2]  Peter J. Rousseeuw,et al.  Robust regression and outlier detection , 1987 .

[3]  W. Härdle,et al.  Robust and Nonlinear Time Series Analysis , 1984 .

[4]  Shinichi Sakata,et al.  S-estimation of nonlinear regression models with dependent and heterogeneous observations , 2001 .

[5]  David E. Tyler,et al.  On the maximum bias functions of MM-estimates and constrained M-estimates of regression , 2007, 0708.0390.

[6]  V. Yohai,et al.  Optimal locally robust M-estimates of regression , 1997 .

[7]  PETER J. ROUSSEEUW,et al.  Computing LTS Regression for Large Data Sets , 2005, Data Mining and Knowledge Discovery.

[8]  Ana Bianco,et al.  Approximate τ—Estimates for Linear Regression Based on Subsampling of Elemental Sets , 2001 .

[9]  V. Yohai,et al.  A Fast Procedure for Outlier Diagnostics in Large Regression Problems , 1999 .

[10]  O. Arslan,et al.  Algorithms to compute CM- and S-estimates for regression , 2002 .

[11]  Cecilia R. Aragon,et al.  Optimization by Simulated Annealing: An Experimental Evaluation; Part I, Graph Partitioning , 1989, Oper. Res..

[12]  V. Yohai,et al.  High Breakdown-Point Estimates of Regression by Means of the Minimization of an Efficient Scale , 1988 .

[13]  Peter J. Rousseeuw,et al.  ROBUST REGRESSION BY MEANS OF S-ESTIMATORS , 1984 .

[14]  V. Yohai HIGH BREAKDOWN-POINT AND HIGH EFFICIENCY ROBUST ESTIMATES FOR REGRESSION , 1987 .

[15]  Fred W. Glover,et al.  Future paths for integer programming and links to artificial intelligence , 1986, Comput. Oper. Res..

[16]  V. Yohai,et al.  A Fast Algorithm for S-Regression Estimates , 2006 .

[17]  V. Yohai,et al.  Robust Statistics: Theory and Methods , 2006 .

[18]  D. Ruppert Computing S Estimators for Regression and Multivariate Location/Dispersion , 1992 .

[19]  P. Rousseeuw Least Median of Squares Regression , 1984 .

[20]  P. Rousseeuw,et al.  Least median of squares estimation in power systems , 1991 .