Difference between revisions of "Mathematical expectation"
m |
Ulf Rehmann (talk | contribs) m (tex done) |
||
Line 1: | Line 1: | ||
− | {{TEX| | + | {{TEX|done}} |
− | + | ||
''mean value, of a random variable'' | ''mean value, of a random variable'' | ||
{{MSC|60-01}} | {{MSC|60-01}} | ||
− | A numerical characteristic of the probability distribution of a random variable. In the most general setting, the mathematical expectation of a random variable | + | A numerical characteristic of the probability distribution of a random variable. In the most general setting, the mathematical expectation of a random variable $ X( \omega ) $, |
+ | $ \omega \in \Omega $, | ||
+ | is defined as the [[Lebesgue integral|Lebesgue integral]] with respect to a [[Probability measure|probability measure]] $ {\mathsf P} $ | ||
+ | on a given [[Probability space|probability space]] $ ( \Omega ,\ {\mathcal A} ,\ {\mathsf P} ) $: | ||
− | + | $$ \tag{* } | |
+ | {\mathsf E} X \ = \ \int\limits _ \Omega X( \omega ) {\mathsf P} (d \omega ), | ||
+ | $$ | ||
− | provided the integral exists. The mathematical expectation of a real-valued random variable may be calculated also as the Lebesgue integral of | + | provided the integral exists. The mathematical expectation of a real-valued random variable may be calculated also as the Lebesgue integral of $ x $ |
+ | with respect to the probability distribution $ {\mathsf P} _ {X} $ | ||
+ | of $ X $: | ||
− | + | $$ | |
+ | {\mathsf E} X \ = \ \int\limits _ {\mathbf R } x {\mathsf P} _ {X} (dx). | ||
+ | $$ | ||
− | The mathematical expectation of a function in | + | The mathematical expectation of a function in $ X $ |
+ | is expressible in terms of the distribution $ {\mathsf P} _ {X} $; | ||
+ | for example, if $ X $ | ||
+ | is a random variable with values in $ \mathbf R $ | ||
+ | and $ f(x) $ | ||
+ | is a single-valued [[Borel function|Borel function]] of $ x $, | ||
+ | then | ||
− | + | $$ | |
+ | {\mathsf E} f(X) \ = \ \int\limits _ \Omega f(X( \omega )) {\mathsf P} (d \omega ) \ = \ | ||
+ | \int\limits _ {\mathbf R ^ {1} } f(x) {\mathsf P} _ {X} (dx). | ||
+ | $$ | ||
− | If | + | If $ F(x) $ |
+ | is the distribution function of $ X $, | ||
+ | then the mathematical expectation of $ X $ | ||
+ | can be represented as the Lebesgue–Stieltjes (or Riemann–Stieltjes) integral | ||
− | + | $$ | |
+ | {\mathsf E} X \ = \ \int\limits _ {- \infty } ^ \infty x \ dF(x); | ||
+ | $$ | ||
− | here integrability of | + | here integrability of $ X $ |
+ | in the sense of (*) is equivalent to the finiteness of the integral | ||
− | + | $$ | |
+ | \int\limits _ {- \infty } ^ \infty x \ dF(x). | ||
+ | $$ | ||
− | In particular cases, if | + | In particular cases, if $ X $ |
+ | has a discrete distribution with possible values $ x _ {k} $, | ||
+ | $ k = 1,\ 2 \dots $ | ||
+ | and corresponding probabilities $ p _ {k} = {\mathsf P} \{ \omega : {X( \omega ) = x _ {k} } \} $, | ||
+ | then | ||
− | + | $$ | |
+ | {\mathsf E} X \ = \ \sum _ { k } x _ {k} p _ {k} ; | ||
+ | $$ | ||
− | if | + | if $ X $ |
+ | has an absolutely continuous distribution with probability density $ p(x) $, | ||
+ | then | ||
− | + | $$ | |
+ | {\mathsf E} X \ = \ \int\limits _ {- \infty } ^ \infty xp(x) \ dx; | ||
+ | $$ | ||
moreover, the existence of the mathematical expectation is equivalent to the absolute convergence of the corresponding series or integral. | moreover, the existence of the mathematical expectation is equivalent to the absolute convergence of the corresponding series or integral. | ||
Line 37: | Line 73: | ||
Main properties of the mathematical expectation: | Main properties of the mathematical expectation: | ||
− | a) | + | a) $ {\mathsf E} X _ {1} \leq {\mathsf E} X _ {2} $ |
+ | whenever $ X _ {1} ( \omega ) \leq X _ {2} ( \omega ) $ | ||
+ | for all $ \omega \in \Omega $; | ||
− | b) | + | b) $ {\mathsf E} C = C $ |
+ | for every real constant $ C $; | ||
− | c) | + | c) $ {\mathsf E} ( \alpha X _ {1} + \beta X _ {2} ) = \alpha {\mathsf E} X _ {1} + \beta {\mathsf E} X _ {2} $ |
+ | for all real $ \alpha $ | ||
+ | and $ \beta $; | ||
− | d) | + | d) $ {\mathsf E} ( \sum _ {n=1} ^ \infty X _ {n} ) = \sum _ {n=1} ^ \infty {\mathsf E} X _ {n} $ |
+ | if the series $ \sum _ {n=1} ^ \infty {\mathsf E} | X _ {n} | $ | ||
+ | converges; | ||
− | e) | + | e) $ g( {\mathsf E} X) \leq {\mathsf E} g(X) $ |
+ | for convex functions $ g $; | ||
f) every bounded random variable has a finite mathematical expectation; | f) every bounded random variable has a finite mathematical expectation; | ||
− | g) | + | g) $ {\mathsf E} ( \prod _ {k=1} ^ {n} X _ {k} ) = \prod _ {k=1} ^ {n} {\mathsf E} X _ {k} $ |
+ | if the random variables $ X _ {1} \dots X _ {k} $ | ||
+ | are mutually independent. | ||
One can naturally define the notion of a random variable with an infinite mathematical expectation. A typical example is provided by the return times in certain random walks (see, e.g., [[Bernoulli random walk|Bernoulli random walk]]). | One can naturally define the notion of a random variable with an infinite mathematical expectation. A typical example is provided by the return times in certain random walks (see, e.g., [[Bernoulli random walk|Bernoulli random walk]]). | ||
Line 55: | Line 101: | ||
The mathematical expectation is used to define many numerical functional characteristics of probability distributions (as the mathematical expectations of appropriate functions in the given random variables), for example, the [[Generating function|generating function]], the [[Characteristic function|characteristic function]] and the moments (cf. [[Moment|Moment]]) of all orders, in particular, the variance (cf. [[Dispersion|Dispersion]]) and the [[Covariance|covariance]]. | The mathematical expectation is used to define many numerical functional characteristics of probability distributions (as the mathematical expectations of appropriate functions in the given random variables), for example, the [[Generating function|generating function]], the [[Characteristic function|characteristic function]] and the moments (cf. [[Moment|Moment]]) of all orders, in particular, the variance (cf. [[Dispersion|Dispersion]]) and the [[Covariance|covariance]]. | ||
− | The mathematical expectation is a characteristic of the location of the values of a random variable (the mean value of its distribution). Here, the mathematical expectation serves as a "typical" value from the distribution and its role is analogous to the role played in mechanics by the statical momentum — the coordinates of the barycentre of a mass distribution. The mathematical expectation differs from other characteristics of location which describe the distribution in general terms — like the median (cf. [[Median (in statistics)|Median (in statistics)]]) and the [[Mode|mode]], by the higher importance that it and its corresponding scatter characteristic, the variance, have in limit theorems of probability theory. The meaning of the mathematical expectation is most completely revealed by the [[Law of large numbers|law of large numbers]] (see also [[Chebyshev inequality in probability theory|Chebyshev inequality in probability theory]]) and the [[Strong law of large numbers|strong law of large numbers]]. In particular, if | + | The mathematical expectation is a characteristic of the location of the values of a random variable (the mean value of its distribution). Here, the mathematical expectation serves as a "typical" value from the distribution and its role is analogous to the role played in mechanics by the statical momentum — the coordinates of the barycentre of a mass distribution. The mathematical expectation differs from other characteristics of location which describe the distribution in general terms — like the median (cf. [[Median (in statistics)|Median (in statistics)]]) and the [[Mode|mode]], by the higher importance that it and its corresponding scatter characteristic, the variance, have in limit theorems of probability theory. The meaning of the mathematical expectation is most completely revealed by the [[Law of large numbers|law of large numbers]] (see also [[Chebyshev inequality in probability theory|Chebyshev inequality in probability theory]]) and the [[Strong law of large numbers|strong law of large numbers]]. In particular, if $ X _ {1} \dots X _ {n} $ |
+ | is a sequence of mutually-independent identically-distributed random variables with finite mathematical expectation $ a = {\mathsf E} X _ {k} $, | ||
+ | then, as $ n \rightarrow \infty $ | ||
+ | and for every $ \epsilon > 0 $, | ||
− | + | $$ | |
+ | {\mathsf P} \left ( \left | | ||
+ | \frac{X _ {1} + \dots + X _ {n} }{n} | ||
+ | -a \right | > \epsilon | ||
+ | \right ) \ \rightarrow \ 0, | ||
+ | $$ | ||
and, in addition, | and, in addition, | ||
− | + | $$ | |
+ | |||
+ | \frac{X _ {1} + \dots + X _ {n} }{n} | ||
+ | \ \rightarrow \ a | ||
+ | $$ | ||
with probability one. | with probability one. | ||
− | The notion of the mathematical expectation as the expected value of a random variable was first noticed in the 18th century in connection with the theory of games of chance. Initially the term "mathematical expectation" was introduced as the expected pay-off of a player, equal to | + | The notion of the mathematical expectation as the expected value of a random variable was first noticed in the 18th century in connection with the theory of games of chance. Initially the term "mathematical expectation" was introduced as the expected pay-off of a player, equal to $ \sum x _ {k} p _ {k} $ |
+ | for possible pay-offs $ x _ {1} \dots x _ {n} $ | ||
+ | with respective probabilities $ p _ {1} \dots p _ {n} $. | ||
+ | Primary contributions in the generalization and utilization of the notion of the mathematical expectation in its contemporary meaning are due to P.L. Chebyshev. | ||
====References==== | ====References==== |
Latest revision as of 11:49, 10 February 2020
mean value, of a random variable
2020 Mathematics Subject Classification: Primary: 60-01 [MSN][ZBL]
A numerical characteristic of the probability distribution of a random variable. In the most general setting, the mathematical expectation of a random variable $ X( \omega ) $, $ \omega \in \Omega $, is defined as the Lebesgue integral with respect to a probability measure $ {\mathsf P} $ on a given probability space $ ( \Omega ,\ {\mathcal A} ,\ {\mathsf P} ) $:
$$ \tag{* } {\mathsf E} X \ = \ \int\limits _ \Omega X( \omega ) {\mathsf P} (d \omega ), $$
provided the integral exists. The mathematical expectation of a real-valued random variable may be calculated also as the Lebesgue integral of $ x $ with respect to the probability distribution $ {\mathsf P} _ {X} $ of $ X $:
$$ {\mathsf E} X \ = \ \int\limits _ {\mathbf R } x {\mathsf P} _ {X} (dx). $$
The mathematical expectation of a function in $ X $ is expressible in terms of the distribution $ {\mathsf P} _ {X} $; for example, if $ X $ is a random variable with values in $ \mathbf R $ and $ f(x) $ is a single-valued Borel function of $ x $, then
$$ {\mathsf E} f(X) \ = \ \int\limits _ \Omega f(X( \omega )) {\mathsf P} (d \omega ) \ = \ \int\limits _ {\mathbf R ^ {1} } f(x) {\mathsf P} _ {X} (dx). $$
If $ F(x) $ is the distribution function of $ X $, then the mathematical expectation of $ X $ can be represented as the Lebesgue–Stieltjes (or Riemann–Stieltjes) integral
$$ {\mathsf E} X \ = \ \int\limits _ {- \infty } ^ \infty x \ dF(x); $$
here integrability of $ X $ in the sense of (*) is equivalent to the finiteness of the integral
$$ \int\limits _ {- \infty } ^ \infty x \ dF(x). $$
In particular cases, if $ X $ has a discrete distribution with possible values $ x _ {k} $, $ k = 1,\ 2 \dots $ and corresponding probabilities $ p _ {k} = {\mathsf P} \{ \omega : {X( \omega ) = x _ {k} } \} $, then
$$ {\mathsf E} X \ = \ \sum _ { k } x _ {k} p _ {k} ; $$
if $ X $ has an absolutely continuous distribution with probability density $ p(x) $, then
$$ {\mathsf E} X \ = \ \int\limits _ {- \infty } ^ \infty xp(x) \ dx; $$
moreover, the existence of the mathematical expectation is equivalent to the absolute convergence of the corresponding series or integral.
Main properties of the mathematical expectation:
a) $ {\mathsf E} X _ {1} \leq {\mathsf E} X _ {2} $ whenever $ X _ {1} ( \omega ) \leq X _ {2} ( \omega ) $ for all $ \omega \in \Omega $;
b) $ {\mathsf E} C = C $ for every real constant $ C $;
c) $ {\mathsf E} ( \alpha X _ {1} + \beta X _ {2} ) = \alpha {\mathsf E} X _ {1} + \beta {\mathsf E} X _ {2} $ for all real $ \alpha $ and $ \beta $;
d) $ {\mathsf E} ( \sum _ {n=1} ^ \infty X _ {n} ) = \sum _ {n=1} ^ \infty {\mathsf E} X _ {n} $ if the series $ \sum _ {n=1} ^ \infty {\mathsf E} | X _ {n} | $ converges;
e) $ g( {\mathsf E} X) \leq {\mathsf E} g(X) $ for convex functions $ g $;
f) every bounded random variable has a finite mathematical expectation;
g) $ {\mathsf E} ( \prod _ {k=1} ^ {n} X _ {k} ) = \prod _ {k=1} ^ {n} {\mathsf E} X _ {k} $ if the random variables $ X _ {1} \dots X _ {k} $ are mutually independent.
One can naturally define the notion of a random variable with an infinite mathematical expectation. A typical example is provided by the return times in certain random walks (see, e.g., Bernoulli random walk).
The mathematical expectation is used to define many numerical functional characteristics of probability distributions (as the mathematical expectations of appropriate functions in the given random variables), for example, the generating function, the characteristic function and the moments (cf. Moment) of all orders, in particular, the variance (cf. Dispersion) and the covariance.
The mathematical expectation is a characteristic of the location of the values of a random variable (the mean value of its distribution). Here, the mathematical expectation serves as a "typical" value from the distribution and its role is analogous to the role played in mechanics by the statical momentum — the coordinates of the barycentre of a mass distribution. The mathematical expectation differs from other characteristics of location which describe the distribution in general terms — like the median (cf. Median (in statistics)) and the mode, by the higher importance that it and its corresponding scatter characteristic, the variance, have in limit theorems of probability theory. The meaning of the mathematical expectation is most completely revealed by the law of large numbers (see also Chebyshev inequality in probability theory) and the strong law of large numbers. In particular, if $ X _ {1} \dots X _ {n} $ is a sequence of mutually-independent identically-distributed random variables with finite mathematical expectation $ a = {\mathsf E} X _ {k} $, then, as $ n \rightarrow \infty $ and for every $ \epsilon > 0 $,
$$ {\mathsf P} \left ( \left | \frac{X _ {1} + \dots + X _ {n} }{n} -a \right | > \epsilon \right ) \ \rightarrow \ 0, $$
and, in addition,
$$ \frac{X _ {1} + \dots + X _ {n} }{n} \ \rightarrow \ a $$
with probability one.
The notion of the mathematical expectation as the expected value of a random variable was first noticed in the 18th century in connection with the theory of games of chance. Initially the term "mathematical expectation" was introduced as the expected pay-off of a player, equal to $ \sum x _ {k} p _ {k} $ for possible pay-offs $ x _ {1} \dots x _ {n} $ with respective probabilities $ p _ {1} \dots p _ {n} $. Primary contributions in the generalization and utilization of the notion of the mathematical expectation in its contemporary meaning are due to P.L. Chebyshev.
References
[K] | A.N. Kolmogorov, "Foundations of the theory of probability" , Chelsea, reprint (1950) (Translated from Russian) MR0032961 |
[F] | W. Feller, "An introduction to probability theory and its applications", 1–2 , Wiley (1957–1971) |
[L] | M. Loève, "Probability theory" , Springer (1978) MR0651017 MR0651018 Zbl 0385.60001 |
[C] | H. Cramér, "Mathematical methods of statistics" , Princeton Univ. Press (1946) MR0016588 Zbl 0063.01014 |
Mathematical expectation. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Mathematical_expectation&oldid=44402