Hypothesis testing: a model selection approach

In many cases when scientists use hypothesis tests they are fully aware that the statistical model itself, and a fortiori the null hypothesis, is only a partial description of reality, and, from a logical point of view, it is not strictly true. In particular, this applies to black-box models—those most frequently used in statistics. From this point of view it does not seem very reasonable to choose between hypotheses in terms of their probability of error or, even worse, their conditional probabilities of error, since these indexes are too rough to be useful. In this paper a wide range of classical hypothesis testing problems are examined from the point of view of model selection: the acceptance or rejection of the null hypothesis of such a problem is considered in terms of the selection of the most appropriate model, when one is nested inside the other. There are several reasons why a (strictly speaking false) model might be developed: it allows us to synthesize data variability, while maintaining a reasonable 2t with observed data, and also it facilitates prediction of new phenomena with a certain degree of accuracy. Quite simply, modeling makes reality more readily understandable. In this context, some problems of the classical hypothesis testing approach are discussed and several alternatives are considered, in developing methods, based on statistical estimation theory, for nested-model selection and a wide class of hypothesis testing problems. These methods are applied to various situations and compared with classical techniques. Finally, a number of questions concerning the nature of statistical inference are posed. c

[1]  José Manuel Corcuera,et al.  INTRINSIC ANALYSIS OF STATISTICAL ESTIMATION , 1995 .

[2]  Irene A. Stegun,et al.  Handbook of Mathematical Functions. , 1966 .

[3]  Jacob Burbea,et al.  THE INFORMATION METRIC FOR UNIVARIATE LINEAR ELLIPTIC MODELS , 1988 .

[4]  J. Burbea Informative Geometry of Probability Spaces , 1984 .

[5]  C. R. Rao,et al.  Entropy differential metric, distance and divergence measures in probability spaces: A unified approach , 1982 .

[6]  J. Srivastava,et al.  Sequential factorial probing designs for identifying and estimating non-negligible factorial effects for the 2m experiment under the tree structure , 1992 .

[7]  R. Royall Statistical Evidence: A Likelihood Paradigm , 1997 .

[8]  O. L. Davies,et al.  Design and analysis of industrial experiments , 1954 .

[9]  N. Čencov Statistical Decision Rules and Optimal Inference , 2000 .

[10]  H. Jeffreys An invariant form for the prior probability in estimation problems , 1946, Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences.

[11]  R. Muirhead Aspects of Multivariate Statistical Theory , 1982, Wiley Series in Probability and Statistics.

[12]  H. Karcher Riemannian center of mass and mollifier smoothing , 1977 .

[13]  J. N. Srivastava,et al.  10 A critique of some aspects of experimental design , 1996, Design and analysis of experiments.

[14]  E. S. Pearson,et al.  On the Problem of the Most Efficient Tests of Statistical Hypotheses , 1933 .

[15]  James O. Berger Statistical Decision Theory , 1980 .

[16]  C. R. Rao,et al.  Information and the Accuracy Attainable in the Estimation of Statistical Parameters , 1992 .

[17]  Jaya Srivastava,et al.  On the inadequacy of the customary orthogonal arrays in quality control and general scientific experimentation, and the need of probing designs of higher revealing power ∗ , 1987 .

[18]  L. Harlow,et al.  What if there were no significance tests , 1997 .

[19]  M. Émery,et al.  Sur le barycentre d'une probabilité dans une variété , 1991 .

[20]  A. W. F. Edwards,et al.  Statistical Evidence: A Likelihood Paradigm. R. Royall, Chapman & Hall, London, 1997. No. of pages: xvi+191. Price: £39.95. ISBN 0-412-04411-0 , 2000 .

[21]  C. Atkinson Rao's distance measure , 1981 .

[22]  Shun-ichi Amari,et al.  Differential-geometrical methods in statistics , 1985 .