If X is an n-by-1 column vector-valued random variable whose expected value is an n-by-1 column vector μ then its variance is the n-by-n nonnegative-definite matrix:
The entries in this matrix are the covariances between the n different scalar components of X. Since the covariance between a scalar-valued random variable and itself is its variance, it follows that in particular the entries on the diagonal of this matrix are the variances of the scalar components of X. This may appear to be a property of this matrix that depends on which coordinate system is chosen on the space in which the random vector X resides. However, it is true generally that if u is any unit vector, then the variance of the projection of X on u is uTΣu. (This point is expanded upon somewhat at [1] (/wiki/Talk:Covariance_matrix). It is a consequence of an identity that appears below.)
Nomenclatures differ. Some statisticians, following the great probabilist William Feller, call this the variance of the random vector X because it is the natural generalization to higher dimensions of the 1-dimensional variance. Other call it the covariance matrix because it is the matrix of covariances between the scalar components of the vector X.
With scalar-valued random variables X we have the identity
This covariance matrix (though very simple) is a very useful tool in many very different areas. From it a transformation matrix[?] can be derived that allows to completly decorrelate the data or from a different point of view to find an optimal basis for representing the data in a compact way. This is called PCA (principal components analysis) in statistics and KL-Transform (Karhunen-Loève transform) in image processing.
Search Encyclopedia
|
Featured Article
|