# Probability distribution

2010 Mathematics Subject Classification: Primary: 60-01 [MSN][ZBL]

One of the basic concepts in probability theory and mathematical statistics. In the modern approach, a suitable probability space $\{\Omega,S,\operatorname P\}$ is taken as a model of a stochastic phenomenon being considered. Here $\Omega$ is a sample space, $S$ is a $\sigma$-algebra of subsets of $\Omega$ specified in some way and $\operatorname P$ is a measure on $S$ such that $\operatorname P(\Omega)=1$ (a probability measure).

Any such measure on $\{\Omega,S\}$ is called a probability distribution (see [K]). But this definition, basic in the axiomatics introduced by A.N. Kolmogorov in 1933, proved to be too general in the course of the further development of the theory and was replaced by more restrictive ones in order to exclude some "pathological" cases. An example was the requirement that the measure $\operatorname P$ be "perfect" (see [GK]). Probability distributions in function spaces are usually required to satisfy some regularity property, usually formulated as separability but also admitting a characterization in different terms (see Separable process and also [P]).

Many of the probability distributions that appear in the specific problems in probability theory and mathematical statistics have been known for a long time and are connected with the basic probability schemes [F]. They are described either by probabilities of discrete values (see Discrete distribution) or by probability densities (see Continuous distribution). There are also tables compiled in certain cases where they are necessary [BS].

Among the basic probability distributions, some are connected with sequences of independent trials (see Binomial distribution; Geometric distribution; Multinomial distribution) and others with the limit laws corresponding to such a probability scheme when the number of trials increases indefinitely (see Normal distribution; Poisson distribution; Arcsine distribution). But these limit distributions may also appear directly in exact form, as in the theory of stochastic processes (see Wiener process; Poisson process), or as solutions of certain equations arising in so-called characterization theorems (see also Normal distribution; Exponential distribution). A uniform distribution, usually considered as a mathematical way of expressing that outcomes of an experiment are equally possible, can also be obtained as a limit distribution (say, by considering sums of large numbers of random variables or some other random variables with sufficiently smooth and "spread out" distributions modulo 1). More probability distributions can be obtained from those mentioned above by means of functional transformations of the corresponding random variables. For example, in mathematical statistics random variables with a normal distribution are used to obtain variables with a "chi-squared" distribution, a non-central "chi-squared" distribution, a Student distribution, a Fisher $F$-distribution, and others.

Important classes of distributions were discovered in connection with asymptotic methods in probability theory and mathematical statistics (see Limit theorems; Stable distribution; Infinitely-divisible distribution; "Omega-squared" distribution).

It is important, both for the theory and in applications, to be able to define a concept of proximity of distributions. The collection of all probability distributions on $\{\Omega,S\}$ can in different ways be turned into a topological space. Weak convergence of probability distributions plays a basic role here (see Distributions, convergence of). In the one-dimensional and finite-dimensional cases the apparatus of characteristic functions (cf. Characteristic function) is a principal instrument for studying convergence of probability distributions.

A complete description of a probability distribution (say, by means of the density of a probability distribution or a distribution function) is often replaced by a limited collection of characteristics. The most widely used of these in the one-dimensional case are the mathematical expectation (the average value), the dispersion (or variance), the median (in statistics), and the moments (cf. Moment). For numerical characteristics of multivariate probability distributions see Correlation (in statistics); Regression.

The statistical analogue of a probability distribution is an empirical distribution. An empirical distribution and its characteristics can be used for the approximate representation of a theoretical distribution and its characteristics (see Statistical estimator). For ways to measure how well an empirical distribution fits a hypothetical one see Statistical hypotheses, verification of; Non-parametric methods in statistics.

#### References

 [K] A.N. Kolmogorov, "Foundations of the theory of probability", Chelsea, reprint (1950) (Translated from Russian) MR0032961 [GK] B.V. Gnedenko, A.N. Kolmogorov, "Limit distributions for sums of independent random variables", Addison-Wesley (1954) (Translated from Russian) MR0062975 Zbl 0056.36001 [P] Yu.V. Prokhorov, "The method of characteristic functionals", Proc. 4-th Berkeley Symp. Math. Stat. Probab., 2, Univ. California Press (1961) pp. 403–419 Zbl 0158.36502 [F] W. Feller, "An introduction to probability theory and its applications", 1–2, Wiley (1957–1971) [BS] L.N. Bol'shev, N.V. Smirnov, "Tables of mathematical statistics", Libr. math. tables, 46, Nauka (1983) (In Russian) (Processed by L.S. Bark and E.S. Kedrova) Zbl 0529.62099 [G] B.V. Gnedenko, "A course of probability theory", Moscow (1969) (In Russian) [C] H. Cramér, "Mathematical methods of statistics", Princeton Univ. Press (1946) MR0016588 Zbl 0063.01014 [N] J. Neveu, "Bases mathématiques du calcul des probabilités", Masson (1970) MR0272004 Zbl 0203.49901

The term "probability distribution" is most frequently applied to the case when $S=\mathbf R$, $\mathbf R^n$ or a space of real functions.