Conditional Mean, Variance, and MMSE

In this blog post, we’ll explore the connections among conditional mean, variance, conditional Probability Density Functions (PDFs), and also its relationship with Minimum Mean Square Error (MMSE) estimator.

Conditional Probability Density Function

Consider two random variables, X and Y (which could be complex). Let fX|Y(x|y) represent the conditional PDF.

Conditioning on a Realization Y=y

When conditioned on a specific realization Y=y (i.e., X|Y=y), the corresponding PDF becomes fX|Y=y(x), dependent on x, its mean, and variance represented as:

Conditioning on a Random Variable Y

When conditioned on the random variable Y (i.e., X|Y), the corresponding PDF becomes fX|Y(x|y), a function of both x and y. Its mean and variance are functions of Y, denoted as E[X|Y] and V[X|Y]. In contrast to the previous case (for a given realization of Y), these terms are random variable. Specifically, each realization of E[X|Y] takes the form E[X|Y=y], and similarly, each realization of V[X|Y] is of the form V[X|Y=y]. In a way, the posterior density of X is different for different realizations of Y=y.

MMSE Estimator

The MMSE estimator is defined as: X^=g(Y)=E[X|Y] Here, g() represents a function. Hence, the MMSE estimator of X is a function of the random variable Y, making the estimator itself a random variable.

For a specific realization Y=y, the estimate is: x^=g(y)=EX[X|Y=y]  This represents the mean of the posterior density fX|Y=y(x).

The MSE for a particular realization Y=y of MMSE estimator is EX[|XX^|2|Y=y]=VX[X|Y=y] This quantity represents the variance of the posterior density fX|Y=y(x).

However, our primary interest lies in the MSE of the estimator across all realizations of Y. Considering this, error variance Vϵ of the MMSE estimator is Vϵ=EX,Y[|XE[X|Y]|2]=EY[EX[|XE[X|Y]|2|Y]]=EY[V[X|Y]]

An intriguing result emerges when exploring the relationship between the expectation of conditional variance and the variance of conditional expectation: V[X]=E[V[X|Y]]+V[E[X|Y]]

Proof:

This result is called as law of total variance. This is similar to law of total expectation EY[EX[X|Y]]=EX[X]. For completion, note that following is true: EY[EX[g(X)|Y]]=EX[g(X)] where g() is any function.

MMSE Estimator for Linear Model

As a supplementary note, when dealing with random vectors (represented by lowercase bold letters for random vectors), consider a scenario with a Gaussian prior on the signal vector xCN(μx,Cx), independent additive Gaussian noise nCN(0,Cn), and a linear receive model incorporating a known channel HCN×K: y=Hx+n

The posterior density of x|y can be expressed as:

Here,

and

Note that in both expressions, the alternative expressions are applicable when covariances of the information signal vector and noise are invertible. Furthermore, we opted to use the notation C to denote the covariance of a matrix, as it is a more prevalent convention, rather than V.

Remarkably, the conditional covariance is independent of y, which is an uncommon property as it typically depends on y. This also implies that the Mean Square Error (MSE) of the MMSE estimator is the same as that of conditional covariance, given by:

Cϵ=Ey[Cx|y]=Cx|y

In conclusion, understanding the interplay between conditional mean, variance, and conditional PDFs provides valuable insights into the MMSE estimator. Exploring these concepts, particularly in scenarios like the MMSE estimator for a linear model with Gaussian prior and independent additive Gaussian noise, sheds light on the relationship between the posterior density, conditional mean, covariance and MSE of MMSE estimator.

Zakir Hussain Shaik
Zakir Hussain Shaik
PhD Student in Communications Systems

My research interests include wireless communications, distributed signal processing, and convex optimization

Related