Encyclopedia > Covariance

Article Content

Covariance

In probability theory and statistics, the covariance between two random variables $X$ and $Y$, with respective expected values $\mu$ and $\nu$ is:

$\operatorname{cov}(X, Y) = E((X - \mu) (Y - \nu))$
This is equivalent to:
$\operatorname{cov}(X, Y) = E(X Y) - \mu \nu$
a formula which is commonly used for calculation.

If $X$ and $Y$ are independent, then their covariance is zero. The converse however is not true. The covariance is sometimes called a measure of "linear dependence" between the two random variables. That phrase does not mean the same thing that is usually means in mathematics, as explicated in the Wikipedia article titled linear dependence, although the meaning is not unrelated. The correlation is a closely related concept used to measure the degree of linear dependence between two variables.

All Wikipedia text is available under the terms of the GNU Free Documentation License

Search Encyclopedia
 Search over one million articles, find something about almost anything!

Featured Article
 DB ... is a disambiguation page; that is, one that just points to other pages that might otherwise have the same name. If you followed a link here, you might wan ...