The Gauss—Markov Theorem and Random Regressors

Abstract In the standard linear regression model with independent, homoscedastic errors, the Gauss—Markov theorem asserts that = (X'X)-1(X'y) is the best linear unbiased estimator of β and, furthermore, that is the best linear unbiased estimator of c'β for all p × 1 vectors c. In the corresponding random regressor model, X is a random sample of size n from a p-variate distribution. If attention is restricted to linear estimators of c'β that are conditionally unbiased, given X, the Gauss—Markov theorem applies. If, however, the estimator is required only to be unconditionally unbiased, the Gauss—Markov theorem may or may not hold, depending on what is known about the distribution of X. The results generalize to the case in which X is a random sample without replacement from a finite population.

[1]  Calyampudi R. Rao,et al.  The theory of least squares when the parameters are stochastic and its application to the analysis of growth curves. , 1965, Biometrika.

[2]  W. R. Buckland,et al.  Distributions in Statistics: Continuous Multivariate Distributions , 1973 .

[3]  David A. Harville,et al.  Extension of the Gauss-Markov Theorem to Include the Estimation of Random Effects , 1976 .

[4]  C. R. Rao,et al.  Linear Statistical Inference and its Applications , 1968 .

[5]  C. Radhakrishna Rao,et al.  Estimation of parameters in the singular Gauss-Markoff model , 1979 .

[6]  D. B. Duncan,et al.  Linear Dynamic Recursive Estimation from the Viewpoint of Regression Analysis , 1972 .

[7]  A. J. Goldman,et al.  Weak generalized inverses and minimum variance linear unbiased estimation , 1964 .

[8]  C. Radhakrishna Rao,et al.  Representations of best linear unbiased estimators in the Gauss-Markoff model with a singular dispersion matrix , 1973 .

[9]  Barr Rosenberg. The Estimation of Stationary Stochastic Regression Parameters Reexamined , 1972 .

[10]  I. Olkin,et al.  Inequalities: Theory of Majorization and Its Applications , 1980 .

[11]  John S. Chipman,et al.  On Least Squares with Insufficient Observations , 1964 .

[12]  E. Lehmann Testing Statistical Hypotheses , 1960 .

[13]  Y. Toyooka AN ITERATED VERSION OF THE GAUSS-MARKOV THEOREM IN GENERALIZED LEAST SQUARES ESTIMATION , 1987 .

[14]  L. Breiman,et al.  Submodel selection and evaluation in regression. The X-random case , 1992 .

[15]  T. Kariya A Nonlinear Version of the Gauss-Markov Theorem , 1985 .

[16]  A. Sarris,et al.  A Bayesian Approach To Estimation Of Time-Varying Regression Coefficients , 1973 .

[17]  E. L. Lehmann,et al.  Theory of point estimation , 1950 .

[18]  Theo K. Dijkstra,et al.  On Model Uncertainty and its Statistical Implications , 1988 .

[19]  Danny Pfeffermann,et al.  On Extensions of the Gauss‐Markov Theorem to the Case of Stochastic Regression Coefficients , 1984 .