Summary
Minimizing\(\smallint \{ \hat \theta (x)\} ^2 f(x)d\mu \) is discussed under the unbiasedness condition:\(\smallint \hat \theta (x)f_i (x)d\mu = c_i (i = 1,...p)\) and the condition (A):f i (x) (i=1, ..., p) are linearly independent\(\smallint \hat \theta (x)f_i (x)d\mu = c_i (i = 1,...p)\), and\({{\{ \sum\limits_{i = 1}^p {a_i f_i (x)^2 } } \mathord{\left/ {\vphantom {{\{ \sum\limits_{i = 1}^p {a_i f_i (x)^2 } } {f(x)d\mu< \infty implies a_{k + 1} = ... = a_p = 0}}} \right. \kern-\nulldelimiterspace} {f(x)d\mu< \infty implies a_{k + 1} = ... = a_p = 0}}\).
Similar content being viewed by others
References
Akahira M, Puri ML, Takeuchi K (1986) Bhattacharyya bound of variances of unbiased estimators in non-regular cases. To appear in the Ann Inst Statist Math 38
Chapman DG, Robbins H (1951) Minimum variance estimation without regularity assumptions. Ann Math Statist 22:581–586
Chatterji SD (1982) A remark on the Cramér-Rao inequality. In: Kallianpur, Krishnaiah, Ghosh (eds) Statistics and probability: Essays in honor of C R Rao. North Holland, New York, pp 193–196
Fend AV (1959) Bounds for the variance of an estimate. Ann Math Statist 30:381–388
Fraser DAS, Guttman I (1952) Bhattacharyya bounds without regularity assumptions. Ann Math Statist 23:629–632
Hammersley JM (1950) On estimating restricted parameters. J Roy Statist Soc (B) 12:192–240
Isii K (1964) Inequalities of the types of Chebyshev and Cramér-Rao and mathematical programming. Ann Inst Statist math 16:277–293
Kiefer J (1952) On minimum variance in non-regular estimation. Ann Math Statist 23:627–629
Sen PK, Ghosh BK (1976) Comparison of some bounds in estimation theory. Ann Statist 4: 755–765; Correction. Ann Statist 5:593
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Takeuchi, K., Akahira, M. A note on minimum variance. Metrika 33, 85–91 (1986). https://doi.org/10.1007/BF01894731
Received:
Issue Date:
DOI: https://doi.org/10.1007/BF01894731