Difference between revisions of "Confidence estimation"
m |
m |
||
Line 1: | Line 1: | ||
− | {{TEX| | + | {{TEX|want}} |
A method in mathematical statistics for the construction of a set of approximate values of the unknown parameters of probability distributions. | A method in mathematical statistics for the construction of a set of approximate values of the unknown parameters of probability distributions. |
Revision as of 10:22, 11 December 2013
A method in mathematical statistics for the construction of a set of approximate values of the unknown parameters of probability distributions.
Let be a random vector assuming values in a set
in a Euclidean space and let the probability distribution of this vector belong to the parametric family of distributions defined by the densities
,
,
, with respect to some measure
. It is assumed that the true value of the parametric point
corresponding to the result of observations of
is unknown. Confidence estimation consists in constructing a certain set
, depending on
, containing the value of a given function
corresponding to the unknown true value of
.
Let be the range of values of the function
,
, and let
,
, be some family of sets belonging to
for all
from
; moreover, it is assumed that for an arbitrary element
and any value of
the probability of the event
is defined. This probability is given by the integral
![]() |
and is said to be the covering probability of the value of by the set
for a given value of
.
If the true value of is unknown, the set
(from the family of sets
,
) which corresponds to the results of observations of
is said to be a confidence set or an interval estimator for the unknown true value of the function
. The confidence probability
, which can be expressed, in terms of the covering probability, by the equation
![]() |
is employed as the probability characteristic of the interval estimator constructed according to the above rule. In other words,
is the probability of covering by the set
of the value of a given function
corresponding to the unknown true parametric point
.
If the confidence probability is independent of
, the interval estimator
is said to be similar to the sampling space. This name is due to the analogy between the formulas
![]() |
and
![]() |
In a more general situation depends on the unknown
, and for this reason the quality of the interval estimator is usually characterized in practical work by the confidence level
![]() |
where the lower bound is calculated over the set . (The confidence level is sometimes called the confidence coefficient.)
Optimization of confidence estimation is defined by the requirements to be met by interval estimators. For instance, if the objective is to construct confidence sets similar to sampling spaces, with a given confidence level (
), the first requirement is expressed by the identity
![]() |
It is natural to look for an interval estimator that covers the true value of with a probability at least that of covering an arbitrary value
. In other words, the second requirement, known as the requirement of unbiasedness, is expressed by the inequality
![]() |
Under these conditions, the "best" interval estimator may reasonably be taken as one which covers any value
other than the true value
with a smaller probability. Hence the third requirement of "highest accuracy" : For any set
other than
and meeting the condition
![]() |
the inequality
![]() |
must be valid.
The task of finding interval estimators satisfying all three requirements is equivalent to the task of constructing unbiased most-powerful statistical tests similar to the sampling space and having significance level
. The problem of the existence of a solution to this problem and its constructive description form the base of the general theory of statistical hypothesis testing.
Confidence estimation is most often used when is a scalar function. Let
,
, be independent random variables subject to the same normal distribution, with the unknown parameters
and
. The problem is to construct an interval estimator for
. Let
![]() |
Since the random variable is subject to the Student distribution with
degrees of freedom, and since this distribution does not depend on the unknown parameters
and
(
,
), it follows that, for any positive
, the probability of occurrence of the event
![]() |
depends only on . If this interval is taken as the interval estimator
for
, it will correspond to the confidence probability
![]() |
which is independent of . Such an interval estimator is known as a confidence interval, while its end points are known as confidence bounds; in this case the confidence interval is a confidence estimator similar to the sampling space. In this example the interval estimator is most accurate unbiased.
References
[1] | S.S. Wilks, "Mathematical statistics" , Wiley (1962) |
[2] | L. Schmetterer, "Introduction to mathematical statistics" , Springer (1974) (Translated from German) |
[3] | E.L. Lehmann, "Testing statistical hypotheses" , Wiley (1986) |
[4] | L.N. Bol'shev, "On the construction of confidence limits" Theor. Probab. Appl. , 10 (1965) pp. 173–177 Teor. Veroyatnost. i Primenen. , 10 : 1 (1965) pp. 187–192 |
Confidence estimation. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Confidence_estimation&oldid=30939