Covariance matrix

From Encyclopedia of Mathematics
Revision as of 17:03, 7 February 2011 by (talk) (Importing text file)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search

The matrix formed from the pairwise covariances of several random variables; more precisely, for the -dimensional vector the covariance matrix is the square matrix , where is the vector of mean values. The components of the covariance matrix are:

and for they are the same as () (that is, the variances of the lie on the principal diagonal). The covariance matrix is a symmetric positive semi-definite matrix. If the covariance matrix is positive definite, then the distribution of is non-degenerate; otherwise it is degenerate. For the random vector the covariance matrix plays the same role as the variance of a random variable. If the variances of the random variables are all equal to 1, then the covariance matrix of is the same as the correlation matrix.

The sample covariance matrix for the sample , where the , , are independent and identically-distributed random -dimensional vectors, consists of the variance and covariance estimators:

where the vector is the arithmetic mean of the . If the are multivariate normally distributed with covariance matrix , then is the maximum-likelihood estimator of ; in this case the joint distribution of the elements of the matrix is called the Wishart distribution; it is one of the fundamental distributions in multivariate statistical analysis by means of which hypotheses concerning the covariance matrix can be tested.

How to Cite This Entry:
Covariance matrix. Encyclopedia of Mathematics. URL:
This article was adapted from an original article by A.V. Prokhorov (originator), which appeared in Encyclopedia of Mathematics - ISBN 1402006098. See original article