A statistical estimator whose expectation is that of the quantity to be estimated. Suppose that in the realization of a random variable taking values in a probability space , , a function has to be estimated, mapping the parameter set into a certain set , and that as an estimator of a statistic is chosen. If is such that
holds for , then is called an unbiased estimator of . An unbiased estimator is frequently called free of systematic errors.
Let be random variables having the same expectation , that is,
In that case the statistic
is an unbiased estimator of . In particular, the arithmetic mean of the observations, , is an unbiased estimator of . In this example .
Let be independent random variables having the same probability law with distribution function , that is,
In this case the empirical distribution function constructed from the observations is an unbiased estimator of , that is, , .
Let be an unbiased estimator of a parameter , that is, , and assume that is a linear function. In that case the statistic is an unbiased estimator of .
The next example shows that there are cases in which unbiased estimators exist and are even unique, but they may turn out to be useless.
Let be a random variable subject to the geometric distribution with parameter of success , that is, for any natural number ,
If is an unbiased estimator of the parameter , it must satisfy the unbiasedness equation , that is,
The unique solution of this equation is
Evidently, is good only when is very close to 1 or 0, otherwise carries no useful information on .
Suppose that a random variable has the binomial law with parameters and , that is, for any ,
It is known that the best unbiased estimator of the parameter (in the sense of minimum quadratic risk) is the statistic . Nevertheless, if is irrational, . This example reflects a general property of random variables that, generally speaking, a random variable need not take values that agree with its expectation. And finally, cases are possible when unbiased estimators do not exist at all. Thus, if under the conditions of Example 5 one takes as the function to be estimated , then (see Example 6) there is no unbiased estimator for .
The preceding examples demonstrate that the concept of an unbiased estimator in its very nature does not necessarily help an experimenter to avoid all the complications that arise in the construction of statistical estimators, since an unbiased estimator may turn out to be very good and even totally useless; it may not be unique or may not exist at all. Moreover, an unbiased estimator, like every point estimator, also has the following deficiency. It only gives an approximate value for the true value of the quantity to be estimated; this quantity was not known before the experiment and remains unknown after it has been performed. So, in the problem of constructing statistical point estimators there is no serious justification for the fact that in all cases they should produce the resulting unbiased estimator, unless it is assumed that the study of unbiased estimators leads to a simple priority theory. For example, the Rao–Cramér inequality has a simple form for unbiased estimators. Namely, if is an unbiased estimator for a function , then under fairly broad conditions of regularity on the family and the function , the Rao–Cramér inequality implies that
where is the Fisher amount of information for . Thus, there is a lower bound for the variance of an unbiased estimator of , namely, . In particular, if , then it follows from (1) that
A statistical estimator for which equality is attained in the Rao–Cramér inequality is called efficient (cf. Efficient estimator). Thus, the statistic in Example 5 is an efficient unbiased estimator of the parameter of the binomial law, since
that is, is the best point estimator of in the sense of minimum quadratic risk in the class of all unbiased estimators.
Naturally, an experimenter is interested in the case when the class of unbiased estimators is rich enough to allow the choice of the best unbiased estimator in some sense. In this context an important role is played by the Rao–Blackwell–Kolmogorov theorem, which allows one to construct an unbiased estimator of minimal variance. This theorem asserts that if the family has a sufficient statistic and is an arbitrary unbiased estimator of a function , then the statistic obtained by averaging over the fixed sufficient statistic has a risk not exceeding that of relative to any convex loss function for all . If the family is complete, the statistic is uniquely determined. That is, the Rao–Blackwell–Kolmogorov theorem implies that unbiased estimators must be looked for in terms of sufficient statistics, if they exist. The practical value of the Rao–Blackwell–Kolmogorov theorem lies in the fact that it gives a recipe for constructing best unbiased estimators, namely: One has to construct an arbitrary unbiased estimator and then average it over a sufficient statistic.
Suppose that a random variable has the Pascal distribution (a negative binomial distribution) with parameters and (, ); that is,
In this case the statistic is an unbiased estimator of . Since is expressed in terms of the sufficient statistic and the system of functions is complete on , is the only unbiased estimator and, consequently, the best estimator of .
Let be a random variable having the binomial law with parameters and . The generating function of this law can be expressed by the formula
which implies that for any integer , the -th derivative
On the other hand,
that is, the statistic
is an unbiased estimator of , and since is expressed in terms of the sufficient statistic and the system of functions is complete on , it follows that is the only, hence the best, unbiased estimator of .
In connection with this example the following question arises: What functions of the parameter admit an unbiased estimator? A.N. Kolmogorov  has shown that this only happens for polynomials of degree . Thus, if
then it follows from (2) that the statistic
is the only unbiased estimator of . This result implies, in particular, that there is no unbiased estimator of .
Let be a random variable subject to the Poisson law with parameter ; that is, for any integer
Since , the observation of by itself is an unbiased estimator of its mathematical expectation . In turn, an unbiased estimator of, say, is . More generally, the statistic
is an unbiased estimator of . This fact implies, in particular, that the statistic
is an unbiased estimator of the function , . Quite generally, if admits an unbiased estimator, then the unbiasedness equation must hold for it, which is equivalent to
From this one deduces that an unbiased estimator exists for any function that admits a power series expansion in its domain of definition .
Suppose that the independent random variables have the same Poisson law with parameter , . The generating function of this law, which can be expressed by the formula
is an entire analytic function and hence has a unique unbiased estimator. In this case a sufficient statistic is , which has the Poisson law with parameter . If is an unbiased estimator of , then it must satisfy the unbiasedness equation
which implies that
that is, an unbiased estimator of the generating function of the Poisson law is the generating function of the binomial law with parameters and .
Examples 6–9 demonstrate that in certain cases, which occur quite frequently in practice, the problem of constructing best estimators is easily solvable, provided that one restricts attention to the class of unbiased estimators. Kolmogorov  has considered the problem of constructing unbiased estimators, in particular, for the distribution function of a normal law with unknown parameters. A more general definition of an unbiased estimator is due to E. Lehmann , according to whom a statistical estimator of a parameter is called unbiased relative to a loss function if
There is also a modification of this definition (see ). Yu.V. Linnik and his students (see ) have established that under fairly wide assumptions the best unbiased estimator is independent of the loss function.
|||A.N. Kolmogorov, "Unbiased estimates" Izv. Akad. Nauk SSSR Ser. Mat. , 14 : 4 (1950) pp. 303–326 (In Russian)|
|||E.L. Lehmann, "Testing statistical hypotheses" , Wiley (1959)|
|||L.B. Klebanov, "A general definition of unbiasedness" Theor. Probab. Appl. , 21 : 3 (1976) pp. 571–585 Teor. Veroyatnost. i. Primenen. , 21 : 3 (1976) pp. 584–598|
|||L.B. Klebanov, Yu.V. Linnik, A.L. Rukhin, "Unbiased estimation and matrix loss functions" Soviet Math. Dokl. , 12 : 5 (1971) pp. 1526–1528 Dokl. Akad. Nauk SSSR , 200 : 5 (1971) pp. 1024–1025|
|||S. Zacks, "The theory of statistical inference" , Wiley (1971)|
Unbiased estimator. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Unbiased_estimator&oldid=11208