Namespaces
Variants
Actions

Difference between revisions of "User:Boris Tsirelson/sandbox"

From Encyclopedia of Mathematics
Jump to: navigation, search
 
(80 intermediate revisions by the same user not shown)
Line 1: Line 1:
In [[probability theory]], a '''standard probability space''' (called also Lebesgue-Rokhlin probability space or just <!-- link to DISAMBIGUATION PAGE, intentionally; please do not "repair" --> [[Lebesgue space]]; the latter term is ambiguous) is a [[probability space]] satisfying certain assumptions introduced by [[Vladimir Rokhlin (Soviet mathematician)|Vladimir Rokhlin]] in 1940. He showed that the [[unit interval]] endowed with the [[Lebesgue measure]] has important advantages over general probability spaces, and can be used as a probability space for all practical purposes in probability theory. The dimension of the unit interval is not a concern, which was clear already to [[Norbert Wiener]]. He constructed the [[Wiener process]] (also called [[Brownian motion]]) in the form of a [[measurable]] [[map (mathematics)|map]] from the unit interval to the [[function space|space of continuous functions]].
 
  
== Short history ==
 
The theory of standard probability spaces was started by [[John von Neumann|von Neumann]] in 1932<ref>{{harv|von Neumann|1932}} and {{harv|Halmos|von Neumann|1942}} are cited in {{harv|Rokhlin|1962|loc=page 2}} and {{harv|Petersen|1983|loc=page 17}}.</ref> and shaped by [[Vladimir Rokhlin (Soviet mathematician)|Vladimir Rokhlin]] in 1940.<ref>Published in short in 1947, in detail in 1949 in Russian and in 1952 in English, reprinted in 1962 {{harv|Rokhlin|1962}}. An unpublished text of 1940 is mentioned in {{harv|Rokhlin|1962|loc=page 2}}. "The theory of Lebesgue spaces in its present form was constructed by V.A. Rokhlin" {{harv|Sinai|1994|loc=page 16}}.</ref> For modernized presentations see {{harv|Haezendonck|1973}}, {{harv|de la Rue|1993}}, {{harv|Itô|1984|loc=Sect. 2.4}} and {{harv|Rudolf|1990|loc=Chapter 2}}.
 
  
Nowadays standard probability spaces may be (and often are) treated in the framework of [[descriptive set theory]], via [[Borel algebra|standard Borel spaces]], see for example {{harv|Kechris|1995|loc=Sect. 17}}. This approach, natural for experts in descriptive set theory, is based on the [[Borel space#Standard Borel spaces and Kuratowski theorems|isomorphism theorem for standard Borel spaces]] {{harv|Kechris|1995|loc=Theorem (15.6)}} whose proof is very difficult for non-experts in descriptive set theory. The original approach of Rokhlin, based on measure theory, leads to much simpler proofs (since measure theory may neglect [[null set]]s, in contrast to descriptive set theory).
 
  
Standard probability spaces are used routinely in [[ergodic theory]],<ref>"In this book we  will deal exclusively with Lebesgue spaces" {{harv|Petersen|1983|loc=page 17}}.</ref><ref>"Ergodic theory on Lebesgue spaces" is the subtitle of the book {{harv|Rudolph|1990}}.</ref> which cannot be said on probability theory. Some probabilists hold the following opinion: only standard probability spaces are pertinent to probability theory, thus, it is a pity that the standardness is not included into the definition of probability space. Others disagree, however.
 
  
Arguments against standardness:
+
{{MSC|62E}}
* the definition of standardness is technically demanding;
+
{{TEX|done}}
* the same about the theorems based on that definition;
 
* it is possible (and natural) to build all the probability theory without the standardness;
 
* [[event (probability theory)|events]] and [[random variables]] are essential, while [[probability space]]s are auxiliary and should not be taken too seriously.
 
  
Arguments in favour of standardness:
+
A
* [[conditioning (probability)|conditioning]] is easy and natural on standard probability spaces, otherwise it becomes obscure;
+
[[Probability distribution|probability distribution]] of a random variable $X$ which takes non-negative integer values, defined by the formula
* the same for [[measure-preserving transformation]]s between probability spaces, [[group actions]] on a probability space, etc.;
+
\begin{equation}\label{*}
* [[ergodic theory]] uses standard probability spaces routinely and successfully;
+
P(X=k)=\frac{ {k+m-1 \choose k}{N-m-k \choose M-m} } { {N \choose M} } \tag{*}
* being unable to eliminate these (auxiliary) probability spaces, we should make them as useful as possible.
+
\end{equation}
 +
where the parameters <math>N,M,m</math> are non-negative integers which satisfy the condition <math>m\leq M\leq N</math>. A negative hypergeometric distribution often arises in a scheme of sampling without replacement. If in the total population of size <math>N</math>, there are <math>M</math>  "marked"  and <math>N-M</math>  "unmarked"  elements, and if the sampling (without replacement) is performed until the number of  "marked"  elements reaches a fixed number <math>m</math>, then the random variable <math>X</math> — the number of  "unmarked"  elements in the sample — has a negative hypergeometric distribution \eqref{*}. The random variable <math>X+m</math> — the size of the sample — also has a negative hypergeometric distribution. The distribution \eqref{*} is called a negative hypergeometric distribution by analogy with the
 +
[[Negative binomial distribution|negative binomial distribution]], which arises in the same way for sampling with replacement.
  
== Definition ==
+
The mathematical expectation and variance of a negative hypergeometric distribution are, respectively, equal to
One of several well-known equivalent definitions of the standardness is given below, after some preparations. All [[probability space]]s are assumed to be [[complete measure|complete]].
 
  
=== Isomorphism ===
+
\begin{equation}
An [[isomorphism]] between two probability spaces <math>\textstyle (\Omega_1,\mathcal{F}_1,P_1) </math>, <math>\textstyle (\Omega_2,\mathcal{F}_2,P_2) </math> is an [[inverse function|invertible]] map <math>\textstyle f : \Omega_1 \to \Omega_2 </math> such that <math>\textstyle f </math> and <math>\textstyle f^{-1} </math> both are (measurable and) [[measure-preserving transformation|measure preserving maps]].
+
m\frac{N-M} {M+1}
 +
\end{equation}
  
Two probability spaces are isomorphic, if there exists an isomorphism between them.
+
and
  
=== Isomorphism modulo zero ===
+
\begin{equation}
Two probability spaces <math>\textstyle (\Omega_1,\mathcal{F}_1,P_1) </math>, <math>\textstyle (\Omega_2,\mathcal{F}_2,P_2) </math> are isomorphic <math>\textstyle \operatorname{mod} \, 0 </math>, if there exist [[null set]]s <math>\textstyle A_1 \subset \Omega_1 </math>, <math>\textstyle A_2 \subset \Omega_2 </math> such that the probability spaces <math>\textstyle \Omega_1 \setminus A_1 </math>, <math>\textstyle \Omega_2 \setminus A_2 </math> are isomorphic (being endowed naturally with sigma-fields and probability measures).
+
m\frac{(N+1)(N-M)} {(M+1)(M+2)}\Big(1-\frac{m}{M+1}\Big) \, .
 +
\end{equation}
  
=== Standard probability space ===
+
When <math>N, M, N-M \to \infty</math> such that <math>M/N\to p</math>, the negative hypergeometric distribution tends to the
A probability space is '''standard''', if it is isomorphic <math>\textstyle \operatorname{mod} \, 0 </math> to an interval with Lebesgue measure, a finite or countable set of atoms, or a combination (disjoint union) of both.
+
[[negative binomial distribution]] with parameters <math>m</math> and <math>p</math>.
  
See {{harv|Rokhlin|1962|loc=Sect. 2.4 (p. 20)}}, {{harv|Haezendonck|1973|loc=Proposition 6 (p. 249) and Remark 2 (p. 250)}}, and {{harv|de la Rue|1993|loc=Theorem 4-3}}. See also {{harv|Kechris|1995|loc=Sect. 17.F}}, and {{harv|Itô|1984|loc=especially Sect. 2.4 and Exercise 3.1(v)}}. In {{harv|Petersen|1983|loc=Definition 4.5 on page 16}} the measure is assumed finite, not necessarily probabilistic. In {{harv|Sinai|1994|loc=Definition 1 on page 16}} atoms are not allowed.
+
The distribution function <math>F(n)</math> of the negative hypergeometric function with parameters <math>N,M,m</math> is related to the
 +
[[Hypergeometric distribution|hypergeometric distribution]] <math>G(m)</math> with parameters <math>N,M,n</math> by the relation
 +
\begin{equation}
 +
F(n) = 1-G(m-1) \, .
 +
\end{equation}
 +
This means that in solving problems in mathematical statistics related to negative hypergeometric distributions, tables of hypergeometric distributions can be used. The negative hypergeometric distribution is used, for example, in
 +
[[Statistical quality control|statistical quality control]].
  
== Examples of non-standard probability spaces ==
+
====References====
=== A naive white noise ===
+
{|
The space of all functions <math>\textstyle f : \mathbb{R} \to \mathbb{R} </math> may be thought of as the product <math>\textstyle \mathbb{R}^\mathbb{R} </math> of a continuum of copies of the real line <math>\textstyle \mathbb{R} </math>. One may endow <math>\textstyle \mathbb{R} </math> with a probability measure, say, the [[standard normal distribution]] <math>\textstyle \gamma = N(0,1) </math>, and treat the space of functions as the product <math>\textstyle (\mathbb{R},\gamma)^\mathbb{R} </math> of a continuum of identical probability spaces <math>\textstyle (\mathbb{R},\gamma) </math>. The [[product measure]] <math>\textstyle \gamma^\mathbb{R} </math> is a probability measure on <math>\textstyle \mathbb{R}^\mathbb{R} </math>. Many non-experts are inclined to believe that <math>\textstyle \gamma^\mathbb{R} </math> describes the so-called [[white noise]].
+
|-
 
+
|valign="top"|{{Ref|Be}}||valign="top"|  Y.K. Belyaev,   "Probability methods of sampling control", Moscow (1975) (In Russian) {{MR|0428663}}
However, it does not. For the white noise, its integral from <math> 0 </math> to <math> 1 </math> should be a random variable distributed <math>\textstyle N(0,1) </math>. In contrast, the integral (from <math> 0 </math> to <math> 1 </math>) of <math>\textstyle f \in \textstyle (\mathbb{R},\gamma)^\mathbb{R} </math> is undefined. Even worse, <math>\textstyle f </math> fails to be [[almost surely]] measurable. Still worse, the probability of <math>\textstyle f </math> being measurable is undefined. And the worst thing: if <math>\textstyle X </math> is a random variable distributed (say) uniformly on <math>\textstyle (0,1) </math> and independent of <math>\textstyle f </math>, then <math>\textstyle f(X) </math> is not a random variable at all! (It lacks measurability.)
+
|-
 
+
|valign="top"|{{Ref|BoSm}}||valign="top"| L.N. Bol'shev,   N.V. Smirnov,   "Tables of mathematical statistics", ''Libr. math. tables'', '''46''', Nauka  (1983) (In Russian) (Processed by L.S. Bark and E.S. Kedrova) {{MR|0243650}} {{ZBL|0529.62099}}
=== A perforated interval ===
 
Let <math>\textstyle Z \subset (0,1) </math> be a set whose [[inner measure|inner]] Lebesgue measure is equal to <math> 0 </math>, but [[outer measure|outer]] Lebesgue measure --- to <math> 1 </math> (thus, <math>\textstyle Z </math> is [[nonmeasurable]] to extreme). There exists a probability measure <math>\textstyle m </math> on <math>\textstyle Z </math> such that <math>\textstyle m(Z \cap A) = \text{mes} (A) </math> for every Lebesgue measurable <math>\textstyle A \subset (0,1) </math>. (Here <math>\textstyle \text{mes}</math> is the Lebesgue measure.) Events and random variables on the probability space <math>\textstyle (Z,m) </math> (treated <math>\textstyle \operatorname{mod} \, 0 </math>) are in a natural one-to-one correspondence with events and random variables on the probability space <math>\textstyle ((0,1),\text{mes}) </math>. Many non-experts are inclined to conclude that the probability space <math>\textstyle (Z,m) </math> is as good as <math>\textstyle ((0,1),\text{mes}) </math>.
 
 
 
However, it is not. A random variable <math>\textstyle X </math> defined by <math>\textstyle X(\omega)=\omega </math> is distributed uniformly on <math>\textstyle (0,1) </math>. The conditional measure, given <math>\textstyle X=x </math>, is just a single atom (at <math>\textstyle x</math>), provided that <math>\textstyle ((0,1),\text{mes}) </math> is the underlying probability space. However, if <math>\textstyle (Z,m) </math> is used instead, then the conditional measure does not exist when <math>\textstyle x \notin Z </math>.
 
 
 
A perforated circle is constructed similarly. Its events and random variables are the same as on the usual circle. The group of rotations acts on them naturally. However, it fails to act on the perforated circle.
 
 
 
See also {{harv|Rudolph|1990|loc=page 17}}.
 
 
 
=== A superfluous measurable set ===
 
Let <math>\textstyle Z \subset (0,1) </math> be as in the previous example. Sets of the form <math>\textstyle ( A \cap Z ) \cup ( B \setminus Z ), </math> where <math>\textstyle A </math> and <math>\textstyle B </math> are arbitrary Lebesgue measurable sets, are a σ-algebra <math>\textstyle \mathcal{F}; </math> it contains the Lebesgue σ-algebra and <math>\textstyle Z. </math> The formula
 
: <math>\displaystyle m \big( ( A \cap Z ) \cup ( B \setminus Z ) \big) = p \, \operatorname{mes} (A) + (1-p) \operatorname{mes} (B) </math>
 
gives the general form of a probability measure <math>\textstyle m </math> on <math>\textstyle \big( (0,1), \mathcal{F} \big) </math> that extends the Lebesgue measure; here <math>\textstyle p \in [0,1] </math> is a parameter. To be specific, we choose <math>\textstyle p = 0.5. </math> Many non-experts are inclined to believe that such an extension of the Lebesgue measure is at least harmless.
 
 
 
However, it is the perforated interval in disguise. The map
 
: <math>\displaystyle f(x) = \begin{cases}
 
0.5 x &\text{for } x \in Z, \\
 
0.5 + 0.5 x &\text{for } x \in (0,1) \setminus Z
 
\end{cases} </math>
 
is an isomorphism between <math>\textstyle \big( (0,1), \mathcal{F}, m \big) </math> and the perforated interval corresponding to the set
 
: <math>\displaystyle Z_1 = \{ 0.5 x : x \in Z \} \cup \{ 0.5 + 0.5 x : x \in (0,1) \setminus Z \} \, ,</math>
 
another set of inner Lebesgue measure 0 but outer Lebesgue measure 1.
 
 
 
See also {{harv|Rudolph|1990|loc=Exercise 2.11 on page 18}}.
 
 
 
== A criterion of standardness ==
 
Standardness of a given probability space <math>\textstyle (\Omega,\mathcal{F},P) </math> is equivalent to a certain property of a measurable map <math>\textstyle f </math> from <math>\textstyle (\Omega,\mathcal{F},P) </math> to a measurable space <math>\textstyle (X,\Sigma). </math> Interestingly, the answer (standard, or not) does not depend on the choice of <math>\textstyle (X,\Sigma) </math> and <math>\textstyle f </math>. This fact is quite useful; one may adapt the choice of <math>\textstyle (X,\Sigma) </math> and <math>\textstyle f </math> to the given <math>\textstyle (\Omega,\mathcal{F},P). </math> No need to examine all cases. It may be convenient to examine a random variable <math>\textstyle f : \Omega \to \mathbb{R}, </math> a random vector <math>\textstyle f : \Omega \to \mathbb{R}^n, </math> a random sequence <math>\textstyle f : \Omega \to \mathbb{R}^\infty, </math> or a sequence of events <math>\textstyle (A_1,A_2,\dots) </math> treates as a sequence of two-valued random variables, <math>\textstyle f : \Omega \to \{0,1\}^\infty. </math>
 
 
 
Two conditions will be imposed on <math>\textstyle f </math> (to be [[injective function|injective]], and generating). Below it is assumed that such <math>\textstyle f </math> is given. The question of its existence will be addressed afterwards.
 
 
 
The probability space <math>\textstyle (\Omega,\mathcal{F},P) </math> is assumed to be [[complete measure|complete]] (otherwise it cannot be standard).
 
 
 
=== A single random variable ===
 
A measurable function <math>\textstyle f : \Omega \to \mathbb{R} </math> induces a [[pushforward measure]], --- the probability measure <math>\textstyle \mu </math> on <math>\textstyle \mathbb{R}, </math> defined by
 
: <math>\displaystyle \mu(B) = P \big( f^{-1}(B) \big) </math> &nbsp;&nbsp; for Borel sets  <math>\textstyle B \subset \mathbb{R}. </math>
 
(It is nothing but the [[probability distribution|distribution]] of the random variable.) The image <math>\textstyle f (\Omega) </math> is always a set of full outer measure,
 
: <math>\displaystyle \mu^* \big( f(\Omega) \big) = 1, </math>
 
but its [[inner measure]] can differ (see ''a perforated interval''). In other words, <math>\textstyle f (\Omega) </math> need not be a set of [[full measure]] <math>\textstyle \mu. </math>
 
 
 
A measurable function <math>\textstyle f : \Omega \to \mathbb{R} </math> is called ''generating'' if <math>\textstyle \mathcal{F} </math> is the completion of the σ-algebra of inverse images <math>\textstyle f^{-1}(B), </math> where <math>\textstyle B \subset \mathbb{R} </math> runs over all Borel sets.
 
 
 
''Caution.'' &nbsp; The following condition is not sufficient for <math>\textstyle f </math> to be generating: for every <math>\textstyle A \in \mathcal{F} </math> there exists a Borel set <math>\textstyle B \subset \mathbb{R} </math> such that <math>\textstyle P ( A \Delta f^{-1}(B) ) = 0. </math> (<math>\textstyle \Delta </math> means [[symmetric difference]]).
 
 
 
'''Theorem.''' Let a measurable function <math>\textstyle f : \Omega \to \mathbb{R} </math> be injective and generating, then the following two conditions are equivalent:
 
* <math>\textstyle f (\Omega) </math> is of full measure <math>\textstyle \mu; </math>
 
* <math> (\Omega,\mathcal{F},P) \,</math> is a standard probability space.
 
 
 
See also {{harv|Itô|1984|loc=Sect. 3.1}}.
 
 
 
=== A random vector ===
 
The same theorem holds for any <math> \mathbb{R}^n \,</math> (in place of <math> \mathbb{R} \,</math>). A measurable function <math> f : \Omega \to \mathbb{R}^n \,</math> may be thought of as a finite sequence of random variables <math> X_1,\dots,X_n : \Omega \to \mathbb{R}, \,</math> and <math> f \,</math> is generating if and only if <math> \mathcal{F} \,</math> is the completion of the σ-algebra generated by <math> X_1,\dots,X_n. \,</math>
 
 
 
=== A random sequence ===
 
The theorem still holds for the space <math> \mathbb{R}^\infty \,</math> of infinite sequences. A measurable function <math> f : \Omega \to \mathbb{R}^\infty \,</math> may be thought of as an infinite sequence of random variables <math> X_1,X_2,\dots : \Omega \to \mathbb{R}, \,</math> and <math> f \,</math> is generating if and only if <math> \mathcal{F} \,</math> is the completion of the σ-algebra generated by <math> X_1,X_2,\dots. \,</math>
 
 
 
=== A sequence of events ===
 
In particular, if the random variables <math> X_n \,</math> take on only two values 0 and 1, we deal with a measurable function <math> f : \Omega \to \{0,1\}^\infty \,</math> and a sequence of sets <math> A_1,A_2,\dots \in \mathcal{F}. \,</math> The function <math> f \,</math> is generating if and only if <math> \mathcal{F} \,</math> is the completion of the σ-algebra generated by <math> A_1,A_2,\dots. \,</math>
 
 
 
In the pioneering work {{harv|Rokhlin|1962}} sequences <math> A_1,A_2,\dots \,</math> that correspond to injective, generating <math> f \,</math> are called ''bases'' of the probability space <math> (\Omega,\mathcal{F},P) \,</math> (see {{harvnb|Rokhlin|1962|loc=Sect. 2.1}}). A basis is called complete mod 0, if <math> f(\Omega) \,</math> is of full measure <math> \mu, \,</math> see {{harv|Rokhlin|1962|loc=Sect. 2.2}}. In the same section Rokhlin proved that if a probability space is complete mod 0 with respect to some basis, then it is complete mod 0 with respect to every other basis, and defines ''Lebesgue spaces'' by this completeness property. See also {{harv|Haezendonck|1973|loc=Prop. 4 and Def. 7}} and {{harv|Rudolph|1990|loc=Sect. 2.3, especially Theorem 2.2}}.
 
 
 
=== Additional remarks ===
 
The four cases treated above are mutually equivalent, and can be united, since the measurable spaces <math> \mathbb{R}, \,</math> <math> \mathbb{R}^n, \,</math> <math> \mathbb{R}^\infty \,</math> and <math> \{0,1\}^\infty \,</math> are mutually isomorphic; they all are [[Borel space#Standard Borel spaces and Kuratowski theorems|standard measurable spaces]] (in other words, standard Borel spaces).
 
 
 
Existence of an injective measurable function from <math>\textstyle (\Omega,\mathcal{F},P) </math> to a standard measurable space <math>\textstyle (X,\Sigma) </math> does not depend on the choice of <math>\textstyle (X,\Sigma). </math> Taking <math>\textstyle (X,\Sigma) =  \{0,1\}^\infty </math> we get the property well-known as being ''countably separated'' (but called ''separable'' in {{harvnb|Itô|1984}}).
 
 
 
Existence of a generating measurable function from <math>\textstyle (\Omega,\mathcal{F},P) </math> to a standard measurable space <math>\textstyle (X,\Sigma) </math> also does not depend on the choice of <math>\textstyle (X,\Sigma). </math> Taking <math>\textstyle (X,\Sigma) =  \{0,1\}^\infty </math> we get the property well-known as being ''countably generated'' (mod 0), see {{harv|Durrett|1996|loc=Exer. I.5}}.
 
{| class="wikitable" style="font-size: 90%; text-align: center; width: auto;"
 
!                        Probability space
 
!                        Countably separated
 
!                        Countably generated
 
!                        Standard
 
 
|-
 
|-
! {{rh}} | Interval with Lebesgue measure
+
|valign="top"|{{Ref|JoKo}}||valign="top"|  N.L. Johnson,  S. Kotz,  "Distributions in statistics, discrete distributions", Wiley  (1969) {{MR|0268996}} {{ZBL|0292.62009}}
| {{yes}}
 
| {{yes}}
 
| {{yes}}
 
 
|-
 
|-
! {{rh}} | Naive white noise
+
|valign="top"|{{Ref|PaJo}}||valign="top"| G.P. Patil,  S.W. Joshi,  "A dictionary and bibliography of discrete distributions", Hafner  (1968) {{MR|0282770}}
| {{no}}
 
| {{no}}
 
| {{no}}
 
 
|-
 
|-
! {{rh}} | Perforated interval
 
| {{yes}}
 
| {{yes}}
 
| {{no}}
 
 
|}
 
|}
 
Every injective measurable function from a ''standard'' probability space to a ''standard'' measurable space is generating. See {{harv|Rokhlin|1962|loc=Sect. 2.5}}, {{harv|Haezendonck|1973|loc=Corollary 2 on page 253}}, {{harv|de la Rue|1993|loc=Theorems 3-4, 3-5}}. This property does not hold for the non-standard probability space dealt with in the subsection "A superfluous measurable set" above.
 
 
''Caution.'' &nbsp; The property of being countably generated is invariant under mod 0 isomorphisms, but the property of being countably separated is not. In fact, a standard probability space <math>\textstyle (\Omega,\mathcal{F},P) </math> is countably separated if and only if the [[cardinality]] of <math>\textstyle \Omega </math> does not exceed [[cardinality of the continuum|continuum]] (see {{harvnb|Itô|1984|loc=Exer. 3.1(v)}}). A standard probability space may contain a null set of any cardinality, thus, it need not be countably separated. However, it always contains a countably separated subset of full measure.
 
 
== Equivalent definitions ==
 
Let <math>\textstyle (\Omega,\mathcal{F},P) </math> be a complete probability space such that the cardinality of <math>\textstyle \Omega </math> does not exceed continuum (the general case is reduced to this special case, see the caution above).
 
 
=== Via absolute measurability ===
 
'''Definition.''' &nbsp; <math>\textstyle (\Omega,\mathcal{F},P) </math> is standard if it is countably separated, countably generated, and absolutely measurable.
 
 
See {{harv|Rokhlin|1962|loc=the end of Sect. 2.3}} and {{harv|Haezendonck|1973|loc=Remark 2 on page 248}}. "Absolutely measurable" means: measurable in every countably separated, countably generated probability space containing it.
 
 
=== Via perfectness ===
 
'''Definition.''' &nbsp; <math>\textstyle (\Omega,\mathcal{F},P) </math> is standard if it is countably separated and perfect.
 
 
See {{harv|Itô|1984|loc=Sect. 3.1}}. "Perfect" means that for every measurable function from <math>\textstyle (\Omega,\mathcal{F},P) </math> to <math> \mathbb{R} \,</math> the image measure is [[regular measure|regular]]. (Here the image measure is defined on all sets whose inverse images belong to <math>\textstyle \mathcal{F} </math>, irrespective of the Borel structure of <math> \mathbb{R} \,</math>).
 
 
=== Via topology ===
 
'''Definition.''' &nbsp; <math>\textstyle (\Omega,\mathcal{F},P) </math> is standard if there exists a [[topological space|topology]] <math>\textstyle \tau </math> on <math>\textstyle \Omega </math> such that
 
* the topological space <math>\textstyle (\Omega,\tau) </math> is [[metrizable]];
 
* <math>\textstyle \mathcal{F} </math> is the completion of the σ-algebra generated by <math>\textstyle \tau </math> (that is, by all open sets);
 
* for every <math>\textstyle \varepsilon > 0 </math> there exists a compact set <math>\textstyle K </math> in <math>\textstyle (\Omega,\tau) </math> such that <math>\textstyle P(K) \ge 1-\varepsilon. </math>
 
 
See {{harv|de la Rue|1993|loc=Sect. 1}}.
 
 
== Verifying the standardness ==
 
Every probability distribution on the space <math>\textstyle \mathbb{R}^n </math> turns it into a standard probability space. (Here, a probability distribution means a probability measure defined initially on the [[Borel sigma-algebra]] and completed.)
 
 
The same holds on every [[Polish space]], see {{harv|Rokhlin|1962|loc=Sect. 2.7 (p. 24)}}, {{harv|Haezendonck|1973|loc=Example 1 (p. 248)}}, {{harv|de la Rue|1993|loc=Theorem 2-3}}, and {{harv|Itô|1984|loc=Theorem 2.4.1}}.
 
 
For example, the Wiener measure turns the Polish space <math>\textstyle C[0,\infty) </math> (of all continuous functions <math>\textstyle [0,\infty) \to \mathbb{R}, </math> endowed with the [[topological space|topology]] of [[local uniform convergence]]) into a standard probability space.
 
 
Another example: for every sequence of random variables, their joint distribution turns the Polish space <math>\textstyle \mathbb{R}^\infty </math> (of sequences; endowed with the [[product topology]]) into a standard probability space.
 
 
(Thus, the idea of [[dimension]], very natural for [[topological space]]s, is utterly inappropriate for standard probability spaces.)
 
 
The [[product measure|product]] of two standard probability spaces is a standard probability space.
 
 
The same holds for the product of countably many spaces, see {{harv|Rokhlin|1962|loc=Sect. 3.4}}, {{harv|Haezendonck|1973|loc=Proposition 12}}, and {{harv|Itô|1984|loc=Theorem 2.4.3}}.
 
 
A measurable subset of a standard probability space is a standard probability space. It is assumed that the set is not a null set, and is endowed with the conditional measure. See {{harv|Rokhlin|1962|loc=Sect. 2.3 (p. 14)}} and {{harv|Haezendonck|1973|loc=Proposition 5}}.
 
 
Every [[probability measure]] on a [[Borel space#Standard Borel spaces and Kuratowski theorems|standard Borel space]] turns it into a standard probability space.
 
 
== Using the standardness ==
 
=== Regular conditional probabilities ===
 
In the discrete setup, the conditional probability is another probability measure, and the conditional expectation may be treated as the (usual) expectation with respect to the conditional measure, see [[conditional expectation]]. In the non-discrete setup, conditioning is often treated indirectly, since the condition may have probability 0, see [[conditional expectation]]. As a result, a number of well-known facts have special 'conditional' counterparts. For example: linearity of the expectation; Jensen's inequality (see [[conditional expectation]]); [[Hölder's inequality]]; the [[monotone convergence theorem#Lebesgue monotone convergence theorem|monotone convergence theorem]], etc.
 
 
Given a random variable <math>\textstyle Y </math> on a probability space <math>\textstyle (\Omega,\mathcal{F},P) </math>, it is natural to try constructing a conditional measure <math>\textstyle P_y </math>, that is, the [[conditional distribution]] of <math>\textstyle \omega \in \Omega </math> given <math>\textstyle Y(\omega)=y </math>. In general this is impossible (see {{harvnb|Durrett|1996|loc=Sect. 4.1(c)}}). However, for a ''standard'' probability space <math>\textstyle (\Omega,\mathcal{F},P) </math> this is possible, and well-known as ''canonical system of measures'' (see {{harvnb|Rokhlin|1962|loc=Sect. 3.1}}), which is basically the same as ''conditional probability measures'' (see {{harvnb|Itô|1984|loc=Sect. 3.5}}), ''disintegration of measure'' (see {{harvnb|Kechris|1995|loc=Exercise (17.35)}}), and [[regular conditional probability|''regular conditional probabilities'']] (see {{harvnb|Durrett|1996|loc=Sect. 4.1(c)}}).
 
 
The conditional Jensen's inequality is just the (usual) Jensen's inequality applied to the conditional measure. The same holds for many other facts.
 
 
=== Measure preserving transformations ===
 
Given two probability spaces <math>\textstyle (\Omega_1,\mathcal{F}_1,P_1) </math>, <math>\textstyle (\Omega_2,\mathcal{F}_2,P_2) </math> and a measure preserving map <math>\textstyle f : \Omega_1 \to \Omega_2 </math>, the image <math>\textstyle f(\Omega_1) </math> need not cover the whole <math>\textstyle    \Omega_2 </math>, it may miss a null set. It may seem that <math>\textstyle P_2(f(\Omega_1)) </math> has to be equal to 1, but it is not so. The outer measure of <math>\textstyle f(\Omega_1) </math> is equal to 1, but the inner measure may differ. However, if the probability spaces <math>\textstyle (\Omega_1,\mathcal{F}_1,P_1) </math>, <math>\textstyle (\Omega_2,\mathcal{F}_2,P_2) </math> are ''standard '' then <math>\textstyle P_2(f(\Omega_1))=1 </math>, see {{harv|de la Rue|1993|loc=Theorem 3-2}}. If <math>\textstyle f </math> is also one-to-one then every <math>\textstyle A \in \mathcal{F}_1 </math> satisfies <math>\textstyle f(A) \in \mathcal{F}_2 </math>, <math>\textstyle P_2(f(A))=P_1(A) </math>. Therefore <math>\textstyle f^{-1} </math> is measurable (and measure preserving). See {{harv|Rokhlin|1962|loc=Sect. 2.5 (p. 20)}} and {{harv|de la Rue|1993|loc=Theorem 3-5}}. See also {{harv|Haezendonck|1973|loc=Proposition 9 (and Remark after it)}}.
 
 
"There is a coherent way to ignore the sets of measure 0 in a measure space" {{harv|Petersen|1983|loc=page 15}}. Striving to get rid of null sets, mathematicians often use equivalence classes of measurable sets or functions. Equivalence classes of measurable subsets of a probability space form a normed [[complete Boolean algebra]] called the ''measure algebra'' (or metric structure). Every measure preserving map <math>\textstyle f : \Omega_1 \to \Omega_2 </math> leads to a homomorphism <math>\textstyle F </math> of measure algebras; basically, <math>\textstyle F(B) = f^{-1}(B) </math> for <math>\textstyle B\in\mathcal{F}_2 </math>.
 
 
It may seem that every homomorphism of measure algebras has to correspond to some measure preserving map, but it is not so. However, for ''standard'' probability spaces each <math>\textstyle F </math> corresponds to some <math>\textstyle f </math>. See {{harv|Rokhlin|1962|loc=Sect. 2.6 (p. 23) and 3.2}}, {{harv|Kechris|1995|loc=Sect. 17.F}}, {{harv|Petersen|1983|loc=Theorem 4.7 on page 17}}.
 
 
==Notes==
 
 
[A] (von Neumann 1932) and (Halmos & von Neumann 1942) are cited in (Rokhlin 1962, page 2) and (Petersen 1983, page 17).
 
 
<references />
 
 
==References==
 
 
[1] Rokhlin, V. A. (1962), "On the fundamental ideas of measure theory", Translations (American Mathematical Society) Series 1 10: 1–54 . Translated from Russian: Рохлин, В. А. (1949), "Об основных понятиях теории меры", Математический Сборник (Новая Серия) 25 (67): 107–150 .
 
 
*{{citation|last=Rokhlin|first=V.A.|author-link=Vladimir Abramovich Rokhlin|year=1962|title=On the fundamental ideas of measure theory|journal=Translations (American Mathematical Society) Series 1|volume=10|pages=1&ndash;54}}. Translated from Russian: {{citation|last=Рохлин|first=В.А.|year=1949|title=Об основных понятиях теории меры|journal=Математический Сборник (новая серия)|volume=25 (67)|pages=107&ndash;150}}.
 
*{{citation|last=von Neumann|first=J.|author-link=John von Neumann|year=1932|title=Einige Sätze über messbare Abbildungen|journal=Annals of Mathematics (2)|volume=33|pages=574&ndash;586}}.
 
*{{citation|last1=Halmos|first1=P.R.|author1-link=Paul Halmos|last2=von Neumann|first2=J.|author2-link=John von Neumann|year=1942|title=Operator methods in classical mechanics, II|journal=Annals of Mathematics (2)|volume=43|pages=332&ndash;350}}.
 
*{{citation|last=Haezendonck|first=J.|year=1973|title=Abstract Lebesgue-Rohlin spaces|journal=Bulletin de la Societe Mathematique de Belgique|volume=25|pages=243&ndash;258}}.
 
*{{citation|last=de la Rue|first=T.|year=1993|contribution=Espaces de Lebesgue|title=Séminaire de Probabilités XXVII|series=Lecture Notes in Mathematics|volume=1557|pages=15&ndash;21|place=Springer, Berlin}}.
 
*{{citation|last=Petersen|first=K.|title=Ergodic theory|year=1983|publisher=Cambridge Univ. Press}}.
 
*{{citation|last=Itô|first=K.|author-link=Kiyoshi Itō|title=Introduction to probability theory|year=1984|publisher=Cambridge Univ. Press}}.
 
*{{citation|last=Rudolph|first=D.J.|title=Fundamentals of measurable dynamics: Ergodic theory on Lebesgue spaces|year=1990|publication-place=Oxford|publisher=Clarendon Press}}.
 
*{{citation|last=Sinai|first=Ya.G.|author-link=Yakov G. Sinai|title=Topics in ergodic theory|year=1994|publisher=Princeton Univ. Press}}.
 
*{{citation|last=Kechris|first=A.S.|author-link=Alexander S. Kechris|title=Classical descriptive set theory|year=1995|publisher=Springer}}.
 
*{{citation|last=Durrett|first=R.|author-link=Rick Durrett|title=Probability: theory and examples|edition=Second|year=1996}}.
 
*{{citation|last=Wiener|first=N.|author-link=Norbert Wiener|title=Nonlinear problems in random theory|year=1958|publisher=M.I.T. Press}}.
 
 
.
 
 
[[:Category:Probability theory]]
 
[[:Category:Measure theory]]
 

Latest revision as of 14:46, 5 June 2017



2020 Mathematics Subject Classification: Primary: 62E [MSN][ZBL]


A probability distribution of a random variable $X$ which takes non-negative integer values, defined by the formula \begin{equation}\label{*} P(X=k)=\frac{ {k+m-1 \choose k}{N-m-k \choose M-m} } { {N \choose M} } \tag{*} \end{equation} where the parameters \(N,M,m\) are non-negative integers which satisfy the condition \(m\leq M\leq N\). A negative hypergeometric distribution often arises in a scheme of sampling without replacement. If in the total population of size \(N\), there are \(M\) "marked" and \(N-M\) "unmarked" elements, and if the sampling (without replacement) is performed until the number of "marked" elements reaches a fixed number \(m\), then the random variable \(X\) — the number of "unmarked" elements in the sample — has a negative hypergeometric distribution \eqref{*}. The random variable \(X+m\) — the size of the sample — also has a negative hypergeometric distribution. The distribution \eqref{*} is called a negative hypergeometric distribution by analogy with the negative binomial distribution, which arises in the same way for sampling with replacement.

The mathematical expectation and variance of a negative hypergeometric distribution are, respectively, equal to

\begin{equation} m\frac{N-M} {M+1} \end{equation}

and

\begin{equation} m\frac{(N+1)(N-M)} {(M+1)(M+2)}\Big(1-\frac{m}{M+1}\Big) \, . \end{equation}

When \(N, M, N-M \to \infty\) such that \(M/N\to p\), the negative hypergeometric distribution tends to the negative binomial distribution with parameters \(m\) and \(p\).

The distribution function \(F(n)\) of the negative hypergeometric function with parameters \(N,M,m\) is related to the hypergeometric distribution \(G(m)\) with parameters \(N,M,n\) by the relation \begin{equation} F(n) = 1-G(m-1) \, . \end{equation} This means that in solving problems in mathematical statistics related to negative hypergeometric distributions, tables of hypergeometric distributions can be used. The negative hypergeometric distribution is used, for example, in statistical quality control.

References

[Be] Y.K. Belyaev, "Probability methods of sampling control", Moscow (1975) (In Russian) MR0428663
[BoSm] L.N. Bol'shev, N.V. Smirnov, "Tables of mathematical statistics", Libr. math. tables, 46, Nauka (1983) (In Russian) (Processed by L.S. Bark and E.S. Kedrova) MR0243650 Zbl 0529.62099
[JoKo] N.L. Johnson, S. Kotz, "Distributions in statistics, discrete distributions", Wiley (1969) MR0268996 Zbl 0292.62009
[PaJo] G.P. Patil, S.W. Joshi, "A dictionary and bibliography of discrete distributions", Hafner (1968) MR0282770
How to Cite This Entry:
Boris Tsirelson/sandbox. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Boris_Tsirelson/sandbox&oldid=19704