next up previous
Next: Karhunen-Loeve Transform (KLT) Up: klt Previous: Multivariate Random Signals

Covariance and Correlation

Let $x_i$ and $x_j$ be two real random variables in a random vector ${\bf x}=[x_1,\cdots,x_N]^T$. The mean and variance of a variable $x_i$ and the covariance and correlation coefficient (normalized correlation) between two variables $x_i$ and $x_j$ are defined below:

Note that the correlation coefficient $r_{ij}=\sigma^2_{i}/\sigma_i\sigma_j$ can be considered as the normalized covariance $\sigma^2_{ij}$.

To obtain these parameters as expectations of the first and second order functions of the random variables, the joint probability density function $p(x_1,\cdots,x_N)$ is required. However, when it is not available, the parameters can still be estimated by averaging the outcomes of a random experiment involving these variables repeated $K$ times:

\begin{displaymath}\hat{\mu}_i=\frac{1}{K}\sum_{k=1}^K x_i^{(k)} \end{displaymath}


\begin{displaymath}
\hat{\sigma^2_i}=\frac{1}{K}\sum_{k=1}^K (x_i^{(k)}-\hat{\mu}_i)^2
=\frac{1}{K}\sum_{k=1}^K x_i^{(k)}^2-\hat{\mu}_i^2
\end{displaymath}


\begin{displaymath}
\hat{\sigma^2_{ij}}=\frac{1}{K}\sum_{k=1}^K (x_i^{(k)}-\hat{...
...c{1}{K}\sum_{k=1}^K x_j^{(k)}x_j^{(k)}-\hat{\mu}_i \hat{\mu}_j
\end{displaymath}


\begin{displaymath}
\hat{r}_{ij}=\frac{\hat{\sigma^2_{ij}}}{\sqrt{ \hat{\sigma_i...
...=\frac{\hat{\sigma^2_{ij}}}{\hat{\sigma_i}\; \hat{\sigma_j} }
\end{displaymath}

To understand intuitively the meaning of these parameters, we consider the following very simple examples.

Examples:

Now we see that the covariance $\sigma_{ij}^2$ represents how much the two ramdom variables $x_i$ and $x_j$ are positively correlated if $\sigma_{ij}^2>0$, negatively correlated if $\sigma_{ij}^2<0$, or not correlated at all if $\sigma_{ij}^2=0$.

Assume a random vector ${\bf x}=[x_1,\cdots,x_N]^T$ is composed of $N$ samples of a signal $x(t)$. The signal samples close to each other tend to be more correlated than those that are farther away, i.e., given $x_i$, we can predict the next sample $x_{i+1}$ with much higher confidence than predicting some $x_j$ which is farther away. Consequently, the elements in the covariance matrix ${\bf\Sigma}_x$ near the main diagonal have higher values than those farther away from the diagonal.


next up previous
Next: Karhunen-Loeve Transform (KLT) Up: klt Previous: Multivariate Random Signals
Ruye Wang 2016-04-06