Robust statistics: a brief introduction and overview

The paper gives a highly condensed first introduction into robust statistics and some guidance for the interpretation of the literature, with some consideration for the uses in geodetics in the background. Robust statistics is the stability theory of statistical procedures. It systematically investigates the effects of deviations from modelling assumptions on known procedures and, if necessary, develops new, better procedures. Common modelling assumptions are those of normality and of independence of the random errors. For both exist sophisticated theories with important practical consequences, and specifically normality can be replaced by any other reasonable parametric model (cf. Huber 1981, Hampel et al. 1986, and also Beran 1994). As a simple example, let us consider n “independent” measurements of the same quantity, for example a distance. In general they will differ, namely by what we now call the “random error”, and the question arises which value we should take as the “best estimate” of the unknown “true value”. This question was already considered by Gauss (1821; cf. also Huber 1972), and he, noticing that he needed the unknown error distribution to answer this question, turned the problem upside down and asked for that error distribution which made a rule “generally accepted as a good one”, namely the arithmetic mean, optimal (in location or shift models). This led to the normal distribution as assumed error distribution, and to a deeper justification (other than by simplicity) of the method of least squares. Less than a hundred years later, almost everybody believed in the “dogma of normality”, the mathematicians because they believed it to be an empirical fact, and the users of statistics because they believed it to be a mathematical theorem. But the central limit theorem, being a limit theorem, only suggests approximate normality under well-specified conditions in real situations; and empirical investigations, already by Bessel (1818) and later by Newcomb (1886), Jeffreys (1939) and others, show that typical error distributions of high-quality data are slightly but clearly longertailed (i.e., with higher kurtosis or standardized 4th moment) than the normal. Gauss (1821) had been careful to talk about “observations of equal accuracy”, but obviously real data have different accuracy, as modeled by Newcomb (1886). The implicit or explicit hope that under approximate (instead of exact) normality least squares would still be approximately optimal a belief still held by many statisticians today was thwarted by Tukey (1960; cf. also Huber 1981 or Hampel et al. 1986) who showed among other things that already under tiny deviations from normality the mean deviation formerly much used as scale measure until the dispute between Fisher (1920) and Eddington was better than the standard deviation, despite its efficiency loss of 12 percent under strict normality. In fact, the (avoidable) efficiency losses of least squares under high-quality data are more typically between 10 percent and 50 percent or even

[1]  R. Fisher A mathematical Examination of the Methods of determining the Accuracy of Observation by the Mean Error, and by the Mean Square Error , 1920 .

[2]  F. Hampel Contributions to the theory of robust estimation , 1968 .

[3]  Jan Beran,et al.  Statistics for long-memory processes , 1994 .

[4]  Werner A. Stahel,et al.  Robust Statistics: The Approach Based on Influence Functions , 1987 .

[5]  Frederick R. Forst,et al.  On robust estimation of the location parameter , 1980 .

[6]  F. Hampel The Breakdown Points of the Mean Combined With Some Rejection Rules , 1985 .

[7]  H. Jeffreys,et al.  The Theory of Probability , 1896 .

[8]  K. F. Gauss,et al.  Theoria combinationis observationum erroribus minimis obnoxiae , 1823 .

[9]  Lamine Mili,et al.  High breakdown point estimation in electric power systems , 1990, IEEE International Symposium on Circuits and Systems.

[10]  P. J. Huber A Robust Version of the Probability Ratio Test , 1965 .

[11]  Daniel A. Relles,et al.  Statisticians are Fairly Robust Estimators of Location , 1977 .

[12]  R. Maronna Robust $M$-Estimators of Multivariate Location and Scatter , 1976 .

[13]  John Law,et al.  Robust Statistics—The Approach Based on Influence Functions , 1986 .

[14]  P. J. Huber Robust Estimation of a Location Parameter , 1964 .

[15]  S. Newcomb A Generalized Theory of the Combination of Observations so as to Obtain the Best Result , 1886 .

[16]  Moti Lal Tiku,et al.  Robust Inference , 1986 .

[17]  Andreas Ruckstuhl,et al.  Partial breakdown in two-factor models , 1997 .

[18]  F. Hampel The Influence Curve and Its Role in Robust Estimation , 1974 .

[19]  W. R. Buckland,et al.  Contributions to Probability and Statistics , 1960 .

[20]  Vic Barnett,et al.  Outliers in Statistical Data , 1980 .

[21]  Howard Wainer,et al.  Robust Regression & Outlier Detection , 1988 .