|
|
(27 intermediate revisions by the same user not shown) |
Line 1: |
Line 1: |
| | | |
− | =Strong Mixing Conditions=
| |
| | | |
− | :Richard C. Bradley
| |
− | :Department of Mathematics, Indiana University, Bloomington, Indiana, USA
| |
| | | |
− | There has been much research on stochastic models
| |
− | that have a well defined, specific structure — for
| |
− | example, [[Markov chain]]s, Gaussian processes, or
| |
− | linear models, including ARMA
| |
− | (autoregressive – moving average) models.
| |
− | However, it became clear in the middle of the last century
| |
− | that there was a need for
| |
− | a theory of statistical inference (e.g. central limit
| |
− | theory) that could be used in the analysis of time series
| |
− | that did not seem to "fit" any such specific structure
| |
− | but which did seem to have some "asymptotic
| |
− | independence" properties.
| |
− | That motivated the development of a broad theory of
| |
− | "strong mixing conditions" to handle such situations.
| |
− | This note is a brief description of that theory.
| |
| | | |
− | The field of strong mixing conditions is a vast area,
| + | {{MSC|62E}} |
− | and a short note such as this cannot even begin to do
| + | {{TEX|done}} |
− | justice to it.
| |
− | Journal articles (with one exception) will not be cited;
| |
− | and many researchers who made important contributions to
| |
− | this field will not be mentioned here.
| |
− | All that can be done here is to give a narrow snapshot
| |
− | of part of the field.
| |
| | | |
− | '''The strong mixing ($\alpha$-mixing) condition.'''
| + | A |
− | Suppose
| + | [[Probability distribution|probability distribution]] of a random variable $X$ which takes non-negative integer values, defined by the formula |
− | $X := (X_k, k \in {\bf Z})$ is a sequence of
| + | \begin{equation}\label{*} |
− | random variables on a given probability space
| + | P(X=k)=\frac{ {k+m-1 \choose k}{N-m-k \choose M-m} } { {N \choose M} } \tag{*} |
− | $(\Omega, {\cal F}, P)$.
| |
− | For $-\infty \leq j \leq \ell \leq \infty$, let
| |
− | ${\cal F}_j^\ell$ denote the $\sigma$-field of events
| |
− | generated by the random variables
| |
− | $X_k,\ j \le k \leq \ell\ (k \in {\bf Z})$.
| |
− | For any two $\sigma$-fields ${\cal A}$ and
| |
− | ${\cal B} \subset {\cal F}$, define the "measure of
| |
− | dependence"
| |
− | \begin{equation} \alpha({\cal A}, {\cal B}) :=
| |
− | \sup_{A \in {\cal A}, B \in {\cal B}}
| |
− | |P(A \cap B) - P(A)P(B)|.
| |
| \end{equation} | | \end{equation} |
− | For the given random sequence $X$, for any positive
| + | where the parameters <math>N,M,m</math> are non-negative integers which satisfy the condition <math>m\leq M\leq N</math>. A negative hypergeometric distribution often arises in a scheme of sampling without replacement. If in the total population of size <math>N</math>, there are <math>M</math> "marked" and <math>N-M</math> "unmarked" elements, and if the sampling (without replacement) is performed until the number of "marked" elements reaches a fixed number <math>m</math>, then the random variable <math>X</math> — the number of "unmarked" elements in the sample — has a negative hypergeometric distribution \eqref{*}. The random variable <math>X+m</math> — the size of the sample — also has a negative hypergeometric distribution. The distribution \eqref{*} is called a negative hypergeometric distribution by analogy with the |
− | integer $n$, define the dependence coefficient
| + | [[Negative binomial distribution|negative binomial distribution]], which arises in the same way for sampling with replacement. |
− | \begin{equation}\alpha(n) = \alpha(X,n) := | + | |
− | \sup_{j \in {\bf Z}}
| + | The mathematical expectation and variance of a negative hypergeometric distribution are, respectively, equal to |
− | \alpha({\cal F}_{-\infty}^j, {\cal F}_{j + n}^{\infty}).
| |
− | \end{equation}
| |
− | By a trivial argument, the sequence of numbers
| |
− | $(\alpha(n), n \in {\bf N})$ is nonincreasing.
| |
− | The random sequence $X$ is said to be "strongly mixing", | |
− | or "$\alpha$-mixing", if $\alpha(n) \to 0$ as
| |
− | $n \to \infty$.
| |
− | This condition was introduced in 1956 by Rosenblatt [Ro1],
| |
− | and was used in that paper in the proof of a central limit
| |
− | theorem.
| |
− | (The phrase "central limit theorem" will henceforth
| |
− | be abbreviated CLT.)
| |
| | | |
− | In the case where the given sequence $X$ is strictly
| + | \begin{equation} |
− | stationary (i.e. its distribution is invariant under a
| + | m\frac{N-M} {M+1} |
− | shift of the indices), eq. (2) also has the simpler form
| |
− | \begin{equation}\alpha(n) = \alpha(X,n) := | |
− | \alpha({\cal F}_{-\infty}^0, {\cal F}_n^{\infty}). | |
| \end{equation} | | \end{equation} |
− | For simplicity, ''in the rest of this note,
| |
− | we shall restrict to strictly stationary sequences.''
| |
− | (Some comments below will have obvious adaptations to
| |
− | nonstationary processes.)
| |
| | | |
− | In particular, for strictly stationary sequences,
| + | and |
− | the strong mixing ($\alpha$-mixing) condition implies Kolmogorov regularity
| |
− | (a trivial "past tail" $\sigma$-field),
| |
− | which in turn implies "mixing" (in the ergodic-theoretic
| |
− | sense), which in turn implies ergodicity.
| |
− | (None of the converse implications holds.)
| |
− | For further related information, see
| |
− | e.g. [Br, v1, Chapter 2].
| |
| | | |
− | '''Comments on limit theory under $\alpha$-mixing.'''
| + | \begin{equation} |
− | Under $\alpha$-mixing and other similar conditions
| + | m\frac{(N+1)(N-M)} {(M+1)(M+2)}\Big(1-\frac{m}{M+1}\Big) \, . |
− | (including ones reviewed below), there has been a vast development of limit theory — for example, | + | \end{equation} |
− | CLTs, weak invariance principles,
| |
− | laws of the iterated logarithm, almost sure invariance
| |
− | principles, and rates of convergence in the strong law of
| |
− | large numbers.
| |
− | For example, the CLT in [Ro1] evolved through
| |
− | subsequent refinements by several researchers
| |
− | into the following "canonical" form.
| |
− | (For its history and a generously detailed presentation | |
− | of its proof, see e.g. [Br, v1,
| |
− | Theorems 1.19 and 10.2].)
| |
| | | |
− | '''Theorem 1.'''
| + | When <math>N, M, N-M \to \infty</math> such that <math>M/N\to p</math>, the negative hypergeometric distribution tends to the |
− | ''Suppose'' $(X_k, k \in {\bf Z})$
| + | [[negative binomial distribution]] with parameters <math>m</math> and <math>p</math>. |
− | ''is a strictly stationary sequence of random variables such that''
| |
− | $EX_0 = 0$, $EX_0^2 < \infty$,
| |
− | $\sigma_n^2 := ES_n^2 \to \infty$ as $n \to \infty$,
| |
− | ''and'' $\alpha(n) \to 0$ ''as'' $n \to \infty$.
| |
− | ''Then the following two conditions (A) and (B) are equivalent:''
| |
| | | |
− | (A) ''The family of random variables''
| + | The distribution function <math>F(n)</math> of the negative hypergeometric function with parameters <math>N,M,m</math> is related to the |
− | $(S_n^2/\sigma_n^2, n \in {\bf N})$ ''is uniformly integrable.''
| + | [[Hypergeometric distribution|hypergeometric distribution]] <math>G(m)</math> with parameters <math>N,M,n</math> by the relation |
− | | + | \begin{equation} |
− | (B) $S_n/\sigma_n \Rightarrow N(0,1)$ ''as''
| + | F(n) = 1-G(m-1) \, . |
− | $n \to \infty$.
| + | \end{equation} |
− | | + | This means that in solving problems in mathematical statistics related to negative hypergeometric distributions, tables of hypergeometric distributions can be used. The negative hypergeometric distribution is used, for example, in |
− | ''If (the hypothesis and) these two equivalent conditions'' (A) ''and'' (B) ''hold, then''
| + | [[Statistical quality control|statistical quality control]]. |
− | $\sigma_n^2 = n \cdot h(n)$ ''for some function'' $h(t),\ t \in (0, \infty)$ ''which is slowly varying as'' $t \to \infty$.
| |
− | | |
− | Here $S_n := X_1 + X_2 + \dots + X_n$; and
| |
− | $\Rightarrow$ denotes convergence in distribution.
| |
− | The assumption $ES_n^2 \to \infty$ is needed here in
| |
− | order to avoid trivial $\alpha$-mixing (or even
| |
− | 1-dependent) counterexamples in which a kind of "cancellation" prevents the partial sums $S_n$ from
| |
− | "growing" (in probability) and becoming asymptotically
| |
− | normal.
| |
− | | |
− | In the context of Theorem 1, if one wants to obtain asymptotic normality of the
| |
− | partial sums (as in condition (B)) without an explicit
| |
− | uniform integrability assumption on the partial sums
| |
− | (as in condition (A)),
| |
− | then as an alternative, one can impose a combination of assumptions on, say, (i) the (marginal) distribution
| |
− | of $X_0$ and (ii) the rate of decay of the
| |
− | numbers $\alpha(n)$ to 0 (the "mixing rate").
| |
− | This involves a "trade-off"; the weaker one assumption
| |
− | is, the stronger the other has to be.
| |
− | One such CLT of Ibragimov in 1962
| |
− | involved such a "trade-off" in which it is assumed that
| |
− | for some $\delta > 0$,
| |
− | $E|X_0|^{2 + \delta} < \infty$ and
| |
− | $\sum_{n=1}^\infty [\alpha(n)]^{\delta/(2 + \delta)}
| |
− | < \infty$.
| |
− | Counterexamples of Davydov in 1973
| |
− | (with just slightly weaker properties) showed that that
| |
− | result is quite sharp.
| |
− | However, it is not at the exact "borderline".
| |
− | From a covariance inequality of Rio in 1993 and a
| |
− | CLT (in fact a weak invariance principle)
| |
− | of Doukhan, Massart, and Rio in 1994, it became clear that
| |
− | the "exact borderline" CLTs of this
| |
− | kind have to involve quantiles of the (marginal)
| |
− | distribution of $X_0$ (rather than just moments). | |
− | For a generously detailed exposition of such CLTs,
| |
− | see [Br, v1, Chapter 10]; and for further
| |
− | related results, see also Rio [Ri].
| |
− | | |
− | Under the hypothesis (first sentence) of Theorem 1
| |
− | (with just finite second moments), | |
− | there is no mixing rate, no matter how fast
| |
− | (short of $m$-dependence), that can insure that
| |
− | a CLT holds.
| |
− | That was shown in 1983 with two different
| |
− | counterexamples, one by the author and the other by
| |
− | Herrndorf.
| |
− | See [Br, v1\&3, Theorem 10.25 and Chapter 31].
| |
− | | |
− | '''Several other classic strong mixing conditions.'''
| |
− | As indicated above, the terms "$\alpha$-mixing" and
| |
− | "strong mixing condition" (singular) both refer to the condition $\alpha(n) \to 0$.
| |
− | (A little caution is in order;
| |
− | in ergodic theory, the term "strong mixing" is often
| |
− | used to refer to the condition of
| |
− | "mixing in the ergodic-theoretic sense",
| |
− | which is weaker than
| |
− | $\alpha$-mixing as noted earlier.)
| |
− | The term "strong mixing conditions" (plural) can
| |
− | reasonably be thought of as referring
| |
− | to all conditions that are at least as strong
| |
− | as (i.e. that imply) $\alpha$-mixing.
| |
− | In the classical theory, five strong mixing conditions
| |
− | (again, plural) have emerged as the most prominent ones:
| |
− | $\alpha$-mixing itself and four others that will be
| |
− | defined here.
| |
− | | |
− | Recall our probability space $(\Omega, {\cal F}, P)$.
| |
− | For any two $\sigma$-fields ${\cal A}$ and
| |
− | ${\cal B} \subset {\cal F}$, define the following four "measures of dependence":
| |
− | \begin{eqnarray}
| |
− | \phi({\cal A}, {\cal B}) &:= &
| |
− | \sup_{A \in {\cal A}, B \in {\cal B}, P(A) > 0}
| |
− | |P(B|A) - P(B)|; \\
| |
− | \psi({\cal A}, {\cal B}) &:= &
| |
− | \sup_{A \in {\cal A}, B \in {\cal B}, P(A) > 0, P(B) > 0}
| |
− | |P(B \cap A)/[P(A)P(B)]\thinspace -\thinspace 1|; \\
| |
− | \rho({\cal A}, {\cal B}) &:= &
| |
− | \sup_{f \in {\cal L}^2({\cal A}),\thinspace g \in {\cal L}^2({\cal B})}
| |
− | |{\rm Corr}(f,g)|; \quad {\rm and} \\
| |
− | \beta ({\cal A}, {\cal B}) &:=& \sup\ (1/2)
| |
− | \sum_{i=1}^I \sum_{j=1}^J |P(A_i \cap B_j) - P(A_i)P(B_j)|
| |
− | \end{eqnarray} | |
− | where the latter supremum is taken over all pairs of finite
| |
− | partitions $(A_1, A_2, \dots, A_I)$ and
| |
− | $(B_1, B_2, \dots, B_J)$ of $\Omega$
| |
− | such that $A_i \in {\cal A}$ for
| |
− | each $i$ and $B_j \in {\cal B}$ for each $j$.
| |
− | In (6), for a given $\sigma$-field
| |
− | ${\cal D} \subset {\cal F}$,
| |
− | the notation ${\cal L}^2({\cal D})$ refers to the space of
| |
− | (equivalence classes of) square-integrable,
| |
− | ${\cal D}$-measurable random variables.
| |
− | | |
− | ==References==
| |
− | | |
− | | |
− | [Br] R.C. Bradley.
| |
− | ''Introduction to Strong Mixing Conditions,''
| |
− | Vols. 1, 2, and 3.
| |
− | Kendrick Press, Heber City (Utah), 2007.
| |
− | | |
− | [DDLLLP] J. Dedecker, P. Doukhan, G. Lang,
| |
− | J.R. León, S. Louhichi, and C. Prieur.
| |
− | ''Weak Dependence: Models, Theory, and Applications.''
| |
− | Lecture Notes in Statistics 190. Springer-Verlag,
| |
− | New York, 2007.
| |
− | | |
− | [DMS] H. Dehling, T. Mikosch, and M. Sørensen,
| |
− | eds.
| |
− | "Empirical Process Techniques for Dependent Data."
| |
− | Birkhäuser, Boston, 2002.
| |
− | | |
− | [De] M. Denker. The central limit theorem for
| |
− | dynamical systems.
| |
− | In: ''Dynamical Systems and Ergodic Theory,''
| |
− | (K. Krzyzewski, ed.), pp. 33-62.
| |
− | Banach Center Publications, Polish Scientific Publishers,
| |
− | Warsaw, 1989.
| |
− | | |
− | [Do] P. Doukhan.
| |
− | ''Mixing: Properties and Examples.''
| |
− | Springer-Verlag, New York, 1995.
| |
− | | |
− | ---------------------------------------------------
| |
− | | |
− | \noindent [HH] P.\ Hall and C.C.\ Heyde.
| |
− | {\it Martingale Limit Theory and its Application\/}.
| |
− | Academic Press, San Diego, 1980.
| |
− | | |
− | \noindent [IR] I.A.\ Ibragimov and Yu.A.\ Rozanov.
| |
− | {\it Gaussian Random Processes\/}.
| |
− | Springer-Verlag, New York, 1978.
| |
− | | |
− | \noindent [Io] M.\ Iosifescu.
| |
− | Doeblin and the metric theory of continued fractions: a
| |
− | functional theoretic solution to Gauss' 1812 problem.
| |
− | In: {\it Doeblin and Modern Probability\/},
| |
− | (H.\ Cohn, ed.), pp.\ 97-110.
| |
− | Contemporary Mathematics 149,
| |
− | American Mathematical Society, Providence, 1993.
| |
− | | |
− | \noindent [Ja] A.\ Jakubowski.
| |
− | {\it Asymptotic Independent Representations for Sums and
| |
− | Order Statistics of Stationary Sequences\/}.
| |
− | Uniwersytet Miko\l aja Kopernika, Toru\'n, Poland, 1991.
| |
− | | |
− | \noindent [LL] Z.\ Lin and C.\ Lu.
| |
− | {\it Limit Theory for Mixing Dependent Random Variables\/}.
| |
− | Kluwer Academic Publishers, Boston, 1996.
| |
− | | |
− | \noindent [LLR] M.R.\ Leadbetter, G.\ Lindgren, and
| |
− | H.\ Rootz\'en.
| |
− | {\it Extremes and Related Properties of Random Sequences
| |
− | and Processes\/}.
| |
− | Springer-Verlag, New York, 1983.
| |
− | | |
− | \noindent [MT] S.P.\ Meyn and R.L.\ Tweedie.
| |
− | {\it Markov Chains and Stochastic Stability\/} (3rd
| |
− | printing). Springer-Verlag, New York, 1996.
| |
− | | |
− | \noindent [Pe] M.\ Peligrad.
| |
− | Conditional central limit theorem via martingale
| |
− | approximation.
| |
− | In: {\it Dependence in Probability, Analysis and Number
| |
− | Theory\/}, (I.\ Berkes, R.C.\ Bradley, H.\ Dehling,
| |
− | M.\ Peligrad, and R.\ Tichy, eds.), pp.\ 295-309.
| |
− | Kendrick Press, Heber City (Utah), 2010.
| |
− | | |
− | \noindent [Ri] E.\ Rio.
| |
− | {\it Th\'eorie Asymptotique des Processus Al\'eatoires Faiblement D\'ependants\/}. \break
| |
− | Math\'ematiques \& Applications 31.
| |
− | Springer, Paris, 2000.
| |
− | | |
− | \noindent [Ro1] M.\ Rosenblatt. A central limit theorem and
| |
− | a strong mixing condition.
| |
− | {\it Proc.\ Natl.\ Acad.\ Sci.\ USA\/} 42 (1956) 43-47.
| |
− | | |
− | \noindent [Ro2] M.\ Rosenblatt.
| |
− | {\it Markov Processes, Structure and Asymptotic Behavior\/}.
| |
− | Springer-Verlag, New York, 1971.
| |
− | | |
− | \noindent [Ro3] M.\ Rosenblatt.
| |
− | {\it Stationary Sequences and Random Fields\/}.
| |
− | Birkh\"auser, Boston, 1985.
| |
| | | |
− | \noindent [\v Zu] I.G.\ \v Zurbenko.
| + | ====References==== |
− | {\it The Spectral Analysis of Time Series\/}. | + | {| |
− | North-Holland, Amsterdam, 1986.
| + | |- |
| + | |valign="top"|{{Ref|Be}}||valign="top"| Y.K. Belyaev, "Probability methods of sampling control", Moscow (1975) (In Russian) {{MR|0428663}} |
| + | |- |
| + | |valign="top"|{{Ref|BoSm}}||valign="top"| L.N. Bol'shev, N.V. Smirnov, "Tables of mathematical statistics", ''Libr. math. tables'', '''46''', Nauka (1983) (In Russian) (Processed by L.S. Bark and E.S. Kedrova) {{MR|0243650}} {{ZBL|0529.62099}} |
| + | |- |
| + | |valign="top"|{{Ref|JoKo}}||valign="top"| N.L. Johnson, S. Kotz, "Distributions in statistics, discrete distributions", Wiley (1969) {{MR|0268996}} {{ZBL|0292.62009}} |
| + | |- |
| + | |valign="top"|{{Ref|PaJo}}||valign="top"| G.P. Patil, S.W. Joshi, "A dictionary and bibliography of discrete distributions", Hafner (1968) {{MR|0282770}} |
| + | |- |
| + | |} |