Abstract
It has long been customary to measure the adequacy of an estimator by the smallness of its mean squared error. The least squares estimators were studied by Gauss and by other authors later in the nineteenth century. A proof that the best unbiased estimator of a linear function of the means of a set of observed random variables is the least squares estimator was given by Markov [12], a modified version of whose proof is given by David and Neyman [4]. A slightly more general theorem is given by Aitken [1]. Fisher [5] indicated that for large samples the maximum likelihood estimator approximately minimizes the mean squared error when compared with other reasonable estimators. This paper will be concerned with optimum properties or failure of optimum properties of the natural estimator in certain special problems with the risk usually measured by the mean squared error or, in the case of several parameters, by a quadratic function of the estimators. We shall first mention some recent papers on this subject and then give some results, mostly unpublished, in greater detail.
This work was supported in part by an ONR contract at Stanford University.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
A.C. Aitken, “On least squares and linear combination of observations,” Proc. Roy. Soc. Edinburgh, Sect. A, Vol. 55 (1935), pp. 42–48.
D. Blackwell, “On the translation parameter problem for discrete variables,” Ann. Math. Statist, Vol. 22 (1951), pp. 393–399.
C. Blyth, “On minimax statistical decision procedures and their admissibility,” Ann. Math. Statist., Vol. 22 (1951), pp. 22–42.
F.N. David and J. Neyman, “Extension of the Markoff theorem of least squares,” Statist. Res. Mem., Vol. 1 (1938), pp. 105–116.
R. A. Fisher, “On the mathematical foundations of theoretical statistics,” Philos. Trans. Roy. Soc. London, Ser. A, Vol. 222 (1922), pp. 309–368.
M.A. Girshick and L.J. Savage, “Bayes and minimax estimates for quadratic loss functions,” Proceedings of the Second Berkeley Symposium on Mathematical Statistics and Probability, Berkeley and Los Angeles, University of California Press, 1951, pp. 53–73.
J.L. Hodges, Jr., and E.L. Lehmann, “Some applications of the Cramér-Rao inequality,” Proceedings of the Second Berkeley Symposium on Mathematical Statistics and Probability, Berkeley and Los Angeles, University of California Press, 1951, pp. 13–22.
S. Karlin, “Admissibility for estimation with quadratic loss,” Ann. Math. Statist., Vol. 29 (1958), pp. 406–436.
J. Kiefer, “Invariance, minimax sequential estimation, and continuous time processes,” Ann. Math. Statist., Vol. 28 (1957), pp. 573–601.
H. Kudo, “On minimax invariant estimators of the transformation parameter,” Nat. Sci. Rep. Ochanomizu Univ., Vol. 6 (1955), pp. 31–73.
E.L. Lehmann, Testing Statistical Hypotheses, New York, Wiley, 1989, pp. 231 and 338.
A. Markov, Calculus of Probability, St. Petersburg, 1908 (2nd ed.). (In Russian.)
E.J.G. Pitman, “Location and scale parameters,” Biometrika, Vol. 30 (1939), pp. 391–421.
C. Stein, “Inadmissibility of the usual estimator for the mean of a multivariate normal distribution,” Proceedings of the Third Berkeley Symposium on Mathematical Statistics and Probability, Berkeley and Los Angeles, University of California Press, 1956, Vol. 1, pp. 197–206.
—, “A necessary and sufficient condition for admissibility,” Ann. Math. Statist, Vol. 26 (1955), pp. 518–522.
—, “The admissibility of Pitman’s estimator for a single location parameter,” Ann. Math. Statist., Vol. 30 (1959), pp. 970–979.
—, “Multiple regression,” Contributions to Probability and Statistics, Essays in Honor of Harold Hotelling, Stanford, Stanford University Press, 1960, pp. 424–443.
A. Wald, “Contributions to the theory of statistical estimation and testing hypotheses,” Ann. Math. Statist., Vol. 10 (1939), pp. 299–326.
R.A. Wijsman, “Random orthogonal transformations and their use in some classical distribution problems in multivariate analysis,” Ann. Math. Statist., Vol. 28 (1957), pp. 415–423.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1992 Springer Science+Business Media New York
About this chapter
Cite this chapter
James, W., Stein, C. (1992). Estimation with Quadratic Loss. In: Kotz, S., Johnson, N.L. (eds) Breakthroughs in Statistics. Springer Series in Statistics. Springer, New York, NY. https://doi.org/10.1007/978-1-4612-0919-5_30
Download citation
DOI: https://doi.org/10.1007/978-1-4612-0919-5_30
Publisher Name: Springer, New York, NY
Print ISBN: 978-0-387-94037-3
Online ISBN: 978-1-4612-0919-5
eBook Packages: Springer Book Archive