Namespaces
Variants
Actions

Difference between revisions of "Mathematical expectation"

From Encyclopedia of Mathematics
Jump to: navigation, search
m
m (tex done)
 
Line 1: Line 1:
{{TEX|want}}
+
{{TEX|done}}
 
+
 
''mean value, of a random variable''
 
''mean value, of a random variable''
  
 
{{MSC|60-01}}
 
{{MSC|60-01}}
  
A numerical characteristic of the probability distribution of a random variable. In the most general setting, the mathematical expectation of a random variable <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m0626301.png" />, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m0626302.png" />, is defined as the [[Lebesgue integral|Lebesgue integral]] with respect to a [[Probability measure|probability measure]] <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m0626303.png" /> on a given [[Probability space|probability space]] <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m0626304.png" />:
+
A numerical characteristic of the probability distribution of a random variable. In the most general setting, the mathematical expectation of a random variable $  X( \omega ) $,  
 +
$  \omega \in \Omega $,  
 +
is defined as the [[Lebesgue integral|Lebesgue integral]] with respect to a [[Probability measure|probability measure]] $  {\mathsf P} $
 +
on a given [[Probability space|probability space]] $  ( \Omega ,\  {\mathcal A} ,\  {\mathsf P} ) $:
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m0626305.png" /></td> <td valign="top" style="width:5%;text-align:right;">(*)</td></tr></table>
+
$$ \tag{* }
 +
{\mathsf E} X \  = \  \int\limits _  \Omega  X( \omega ) {\mathsf P} (d \omega ),
 +
$$
  
provided the integral exists. The mathematical expectation of a real-valued random variable may be calculated also as the Lebesgue integral of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m0626306.png" /> with respect to the probability distribution <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m0626307.png" /> of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m0626308.png" />:
+
provided the integral exists. The mathematical expectation of a real-valued random variable may be calculated also as the Lebesgue integral of $  x $
 +
with respect to the probability distribution $  {\mathsf P} _ {X} $
 +
of $  X $:
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m0626309.png" /></td> </tr></table>
+
$$
 +
{\mathsf E} X \  = \  \int\limits _ {\mathbf R } x {\mathsf P} _ {X} (dx).
 +
$$
  
The mathematical expectation of a function in <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263010.png" /> is expressible in terms of the distribution <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263011.png" />; for example, if <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263012.png" /> is a random variable with values in <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263013.png" /> and <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263014.png" /> is a single-valued [[Borel function|Borel function]] of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263015.png" />, then
+
The mathematical expectation of a function in $  X $
 +
is expressible in terms of the distribution $  {\mathsf P} _ {X} $;  
 +
for example, if $  X $
 +
is a random variable with values in $  \mathbf R $
 +
and $  f(x) $
 +
is a single-valued [[Borel function|Borel function]] of $  x $,  
 +
then
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263016.png" /></td> </tr></table>
+
$$
 +
{\mathsf E} f(X) \  = \  \int\limits _  \Omega  f(X( \omega )) {\mathsf P} (d \omega ) \  = \
 +
\int\limits _ {\mathbf R  ^ {1} } f(x) {\mathsf P} _ {X} (dx).
 +
$$
  
If <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263017.png" /> is the distribution function of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263018.png" />, then the mathematical expectation of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263019.png" /> can be represented as the Lebesgue–Stieltjes (or Riemann–Stieltjes) integral
+
If $  F(x) $
 +
is the distribution function of $  X $,  
 +
then the mathematical expectation of $  X $
 +
can be represented as the Lebesgue–Stieltjes (or Riemann–Stieltjes) integral
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263020.png" /></td> </tr></table>
+
$$
 +
{\mathsf E} X \  = \  \int\limits _ {- \infty } ^  \infty  x \  dF(x);
 +
$$
  
here integrability of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263021.png" /> in the sense of (*) is equivalent to the finiteness of the integral
+
here integrability of $  X $
 +
in the sense of (*) is equivalent to the finiteness of the integral
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263022.png" /></td> </tr></table>
+
$$
 +
\int\limits _ {- \infty } ^  \infty  x \  dF(x).
 +
$$
  
In particular cases, if <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263023.png" /> has a discrete distribution with possible values <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263024.png" />, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263025.png" /> and corresponding probabilities <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263026.png" />, then
+
In particular cases, if $  X $
 +
has a discrete distribution with possible values $  x _ {k} $,
 +
$  k = 1,\  2 \dots $
 +
and corresponding probabilities $  p _ {k} = {\mathsf P} \{  \omega  : {X( \omega ) = x _ {k} } \} $,  
 +
then
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263027.png" /></td> </tr></table>
+
$$
 +
{\mathsf E} X \  = \  \sum _ { k } x _ {k} p _ {k} ;
 +
$$
  
if <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263028.png" /> has an absolutely continuous distribution with probability density <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263029.png" />, then
+
if $  X $
 +
has an absolutely continuous distribution with probability density $  p(x) $,  
 +
then
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263030.png" /></td> </tr></table>
+
$$
 +
{\mathsf E} X \  = \  \int\limits _ {- \infty } ^  \infty  xp(x) \  dx;
 +
$$
  
 
moreover, the existence of the mathematical expectation is equivalent to the absolute convergence of the corresponding series or integral.
 
moreover, the existence of the mathematical expectation is equivalent to the absolute convergence of the corresponding series or integral.
Line 37: Line 73:
 
Main properties of the mathematical expectation:
 
Main properties of the mathematical expectation:
  
a) <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263031.png" /> whenever <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263032.png" /> for all <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263033.png" />;
+
a) $  {\mathsf E} X _ {1} \leq  {\mathsf E} X _ {2} $
 +
whenever $  X _ {1} ( \omega ) \leq  X _ {2} ( \omega ) $
 +
for all $  \omega \in \Omega $;
  
b) <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263034.png" /> for every real constant <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263035.png" />;
+
b) $  {\mathsf E} C = C $
 +
for every real constant $  C $;
  
c) <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263036.png" /> for all real <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263037.png" /> and <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263038.png" />;
+
c) $  {\mathsf E} ( \alpha X _ {1} + \beta X _ {2} ) = \alpha {\mathsf E} X _ {1} + \beta {\mathsf E} X _ {2} $
 +
for all real $  \alpha $
 +
and $  \beta $;
  
d) <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263039.png" /> if the series <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263040.png" /> converges;
+
d) $  {\mathsf E} ( \sum _ {n=1}  ^  \infty  X _ {n} ) = \sum _ {n=1}  ^  \infty  {\mathsf E} X _ {n} $
 +
if the series $  \sum _ {n=1}  ^  \infty  {\mathsf E} | X _ {n} | $
 +
converges;
  
e) <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263041.png" /> for convex functions <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263042.png" />;
+
e) $  g( {\mathsf E} X) \leq  {\mathsf E} g(X) $
 +
for convex functions $  g $;
  
 
f) every bounded random variable has a finite mathematical expectation;
 
f) every bounded random variable has a finite mathematical expectation;
  
g) <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263043.png" /> if the random variables <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263044.png" /> are mutually independent.
+
g) $  {\mathsf E} ( \prod _ {k=1}  ^ {n} X _ {k} ) = \prod _ {k=1}  ^ {n} {\mathsf E} X _ {k} $
 +
if the random variables $  X _ {1} \dots X _ {k} $
 +
are mutually independent.
  
 
One can naturally define the notion of a random variable with an infinite mathematical expectation. A typical example is provided by the return times in certain random walks (see, e.g., [[Bernoulli random walk|Bernoulli random walk]]).
 
One can naturally define the notion of a random variable with an infinite mathematical expectation. A typical example is provided by the return times in certain random walks (see, e.g., [[Bernoulli random walk|Bernoulli random walk]]).
Line 55: Line 101:
 
The mathematical expectation is used to define many numerical functional characteristics of probability distributions (as the mathematical expectations of appropriate functions in the given random variables), for example, the [[Generating function|generating function]], the [[Characteristic function|characteristic function]] and the moments (cf. [[Moment|Moment]]) of all orders, in particular, the variance (cf. [[Dispersion|Dispersion]]) and the [[Covariance|covariance]].
 
The mathematical expectation is used to define many numerical functional characteristics of probability distributions (as the mathematical expectations of appropriate functions in the given random variables), for example, the [[Generating function|generating function]], the [[Characteristic function|characteristic function]] and the moments (cf. [[Moment|Moment]]) of all orders, in particular, the variance (cf. [[Dispersion|Dispersion]]) and the [[Covariance|covariance]].
  
The mathematical expectation is a characteristic of the location of the values of a random variable (the mean value of its distribution). Here, the mathematical expectation serves as a "typical" value from the distribution and its role is analogous to the role played in mechanics by the statical momentum — the coordinates of the barycentre of a mass distribution. The mathematical expectation differs from other characteristics of location which describe the distribution in general terms — like the median (cf. [[Median (in statistics)|Median (in statistics)]]) and the [[Mode|mode]], by the higher importance that it and its corresponding scatter characteristic, the variance, have in limit theorems of probability theory. The meaning of the mathematical expectation is most completely revealed by the [[Law of large numbers|law of large numbers]] (see also [[Chebyshev inequality in probability theory|Chebyshev inequality in probability theory]]) and the [[Strong law of large numbers|strong law of large numbers]]. In particular, if <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263045.png" /> is a sequence of mutually-independent identically-distributed random variables with finite mathematical expectation <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263046.png" />, then, as <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263047.png" /> and for every <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263048.png" />,
+
The mathematical expectation is a characteristic of the location of the values of a random variable (the mean value of its distribution). Here, the mathematical expectation serves as a "typical" value from the distribution and its role is analogous to the role played in mechanics by the statical momentum — the coordinates of the barycentre of a mass distribution. The mathematical expectation differs from other characteristics of location which describe the distribution in general terms — like the median (cf. [[Median (in statistics)|Median (in statistics)]]) and the [[Mode|mode]], by the higher importance that it and its corresponding scatter characteristic, the variance, have in limit theorems of probability theory. The meaning of the mathematical expectation is most completely revealed by the [[Law of large numbers|law of large numbers]] (see also [[Chebyshev inequality in probability theory|Chebyshev inequality in probability theory]]) and the [[Strong law of large numbers|strong law of large numbers]]. In particular, if $  X _ {1} \dots X _ {n} $
 +
is a sequence of mutually-independent identically-distributed random variables with finite mathematical expectation $  a = {\mathsf E} X _ {k} $,  
 +
then, as $  n \rightarrow \infty $
 +
and for every $  \epsilon > 0 $,
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263049.png" /></td> </tr></table>
+
$$
 +
{\mathsf P} \left ( \left |
 +
\frac{X _ {1} + \dots + X _ {n} }{n}
 +
-a \right | > \epsilon
 +
\right ) \  \rightarrow \  0,
 +
$$
  
 
and, in addition,
 
and, in addition,
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263050.png" /></td> </tr></table>
+
$$
 +
 
 +
\frac{X _ {1} + \dots + X _ {n} }{n}
 +
\  \rightarrow \  a
 +
$$
  
 
with probability one.
 
with probability one.
  
The notion of the mathematical expectation as the expected value of a random variable was first noticed in the 18th century in connection with the theory of games of chance. Initially the term "mathematical expectation" was introduced as the expected pay-off of a player, equal to <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263051.png" /> for possible pay-offs <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263052.png" /> with respective probabilities <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m062/m062630/m06263053.png" />. Primary contributions in the generalization and utilization of the notion of the mathematical expectation in its contemporary meaning are due to P.L. Chebyshev.
+
The notion of the mathematical expectation as the expected value of a random variable was first noticed in the 18th century in connection with the theory of games of chance. Initially the term "mathematical expectation" was introduced as the expected pay-off of a player, equal to $  \sum x _ {k} p _ {k} $
 +
for possible pay-offs $  x _ {1} \dots x _ {n} $
 +
with respective probabilities $  p _ {1} \dots p _ {n} $.  
 +
Primary contributions in the generalization and utilization of the notion of the mathematical expectation in its contemporary meaning are due to P.L. Chebyshev.
  
 
====References====
 
====References====

Latest revision as of 11:49, 10 February 2020


mean value, of a random variable

2020 Mathematics Subject Classification: Primary: 60-01 [MSN][ZBL]

A numerical characteristic of the probability distribution of a random variable. In the most general setting, the mathematical expectation of a random variable $ X( \omega ) $, $ \omega \in \Omega $, is defined as the Lebesgue integral with respect to a probability measure $ {\mathsf P} $ on a given probability space $ ( \Omega ,\ {\mathcal A} ,\ {\mathsf P} ) $:

$$ \tag{* } {\mathsf E} X \ = \ \int\limits _ \Omega X( \omega ) {\mathsf P} (d \omega ), $$

provided the integral exists. The mathematical expectation of a real-valued random variable may be calculated also as the Lebesgue integral of $ x $ with respect to the probability distribution $ {\mathsf P} _ {X} $ of $ X $:

$$ {\mathsf E} X \ = \ \int\limits _ {\mathbf R } x {\mathsf P} _ {X} (dx). $$

The mathematical expectation of a function in $ X $ is expressible in terms of the distribution $ {\mathsf P} _ {X} $; for example, if $ X $ is a random variable with values in $ \mathbf R $ and $ f(x) $ is a single-valued Borel function of $ x $, then

$$ {\mathsf E} f(X) \ = \ \int\limits _ \Omega f(X( \omega )) {\mathsf P} (d \omega ) \ = \ \int\limits _ {\mathbf R ^ {1} } f(x) {\mathsf P} _ {X} (dx). $$

If $ F(x) $ is the distribution function of $ X $, then the mathematical expectation of $ X $ can be represented as the Lebesgue–Stieltjes (or Riemann–Stieltjes) integral

$$ {\mathsf E} X \ = \ \int\limits _ {- \infty } ^ \infty x \ dF(x); $$

here integrability of $ X $ in the sense of (*) is equivalent to the finiteness of the integral

$$ \int\limits _ {- \infty } ^ \infty x \ dF(x). $$

In particular cases, if $ X $ has a discrete distribution with possible values $ x _ {k} $, $ k = 1,\ 2 \dots $ and corresponding probabilities $ p _ {k} = {\mathsf P} \{ \omega : {X( \omega ) = x _ {k} } \} $, then

$$ {\mathsf E} X \ = \ \sum _ { k } x _ {k} p _ {k} ; $$

if $ X $ has an absolutely continuous distribution with probability density $ p(x) $, then

$$ {\mathsf E} X \ = \ \int\limits _ {- \infty } ^ \infty xp(x) \ dx; $$

moreover, the existence of the mathematical expectation is equivalent to the absolute convergence of the corresponding series or integral.

Main properties of the mathematical expectation:

a) $ {\mathsf E} X _ {1} \leq {\mathsf E} X _ {2} $ whenever $ X _ {1} ( \omega ) \leq X _ {2} ( \omega ) $ for all $ \omega \in \Omega $;

b) $ {\mathsf E} C = C $ for every real constant $ C $;

c) $ {\mathsf E} ( \alpha X _ {1} + \beta X _ {2} ) = \alpha {\mathsf E} X _ {1} + \beta {\mathsf E} X _ {2} $ for all real $ \alpha $ and $ \beta $;

d) $ {\mathsf E} ( \sum _ {n=1} ^ \infty X _ {n} ) = \sum _ {n=1} ^ \infty {\mathsf E} X _ {n} $ if the series $ \sum _ {n=1} ^ \infty {\mathsf E} | X _ {n} | $ converges;

e) $ g( {\mathsf E} X) \leq {\mathsf E} g(X) $ for convex functions $ g $;

f) every bounded random variable has a finite mathematical expectation;

g) $ {\mathsf E} ( \prod _ {k=1} ^ {n} X _ {k} ) = \prod _ {k=1} ^ {n} {\mathsf E} X _ {k} $ if the random variables $ X _ {1} \dots X _ {k} $ are mutually independent.

One can naturally define the notion of a random variable with an infinite mathematical expectation. A typical example is provided by the return times in certain random walks (see, e.g., Bernoulli random walk).

The mathematical expectation is used to define many numerical functional characteristics of probability distributions (as the mathematical expectations of appropriate functions in the given random variables), for example, the generating function, the characteristic function and the moments (cf. Moment) of all orders, in particular, the variance (cf. Dispersion) and the covariance.

The mathematical expectation is a characteristic of the location of the values of a random variable (the mean value of its distribution). Here, the mathematical expectation serves as a "typical" value from the distribution and its role is analogous to the role played in mechanics by the statical momentum — the coordinates of the barycentre of a mass distribution. The mathematical expectation differs from other characteristics of location which describe the distribution in general terms — like the median (cf. Median (in statistics)) and the mode, by the higher importance that it and its corresponding scatter characteristic, the variance, have in limit theorems of probability theory. The meaning of the mathematical expectation is most completely revealed by the law of large numbers (see also Chebyshev inequality in probability theory) and the strong law of large numbers. In particular, if $ X _ {1} \dots X _ {n} $ is a sequence of mutually-independent identically-distributed random variables with finite mathematical expectation $ a = {\mathsf E} X _ {k} $, then, as $ n \rightarrow \infty $ and for every $ \epsilon > 0 $,

$$ {\mathsf P} \left ( \left | \frac{X _ {1} + \dots + X _ {n} }{n} -a \right | > \epsilon \right ) \ \rightarrow \ 0, $$

and, in addition,

$$ \frac{X _ {1} + \dots + X _ {n} }{n} \ \rightarrow \ a $$

with probability one.

The notion of the mathematical expectation as the expected value of a random variable was first noticed in the 18th century in connection with the theory of games of chance. Initially the term "mathematical expectation" was introduced as the expected pay-off of a player, equal to $ \sum x _ {k} p _ {k} $ for possible pay-offs $ x _ {1} \dots x _ {n} $ with respective probabilities $ p _ {1} \dots p _ {n} $. Primary contributions in the generalization and utilization of the notion of the mathematical expectation in its contemporary meaning are due to P.L. Chebyshev.

References

[K] A.N. Kolmogorov, "Foundations of the theory of probability" , Chelsea, reprint (1950) (Translated from Russian) MR0032961
[F] W. Feller, "An introduction to probability theory and its applications", 1–2 , Wiley (1957–1971)
[L] M. Loève, "Probability theory" , Springer (1978) MR0651017 MR0651018 Zbl 0385.60001
[C] H. Cramér, "Mathematical methods of statistics" , Princeton Univ. Press (1946) MR0016588 Zbl 0063.01014
How to Cite This Entry:
Mathematical expectation. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Mathematical_expectation&oldid=29499
This article was adapted from an original article by A.V. Prokhorov (originator), which appeared in Encyclopedia of Mathematics - ISBN 1402006098. See original article