Multivariate normal distribution is normalized (of course): A proof



Let X=(X1,X2,,Xn) be a vector of random variables. We say it follows the multivariate normal (Gaussian) distribution if its density is given by

(Eq:density)f(x)=1(2π)n|Σ|exp(12(xμ)Σ1(xμ))

where μ=(μ1,μ2,,μn)Rn is a vector, Σ is a symmetric positive definite n×n matrix, and Σ1 and |Σ| are the inverse and determinant of Σ, respectively. It turns out that μ and Σ are the mean vector and covariance matrix of X, respectively. But we will not prove that here. In this post, we will show this density (Eq:density) is normalized (of course). That is, we prove that

Rnf(x)dx=1.

We assume that you already know how to prove the univariate normal distribution is normalized:

12πσ2exp((xμ)22σ2)dx=1.

Let's start!

First, by changing the variables y=xμ, we need to prove

(Eq:Goal)Rnexp(12yΣ1y)dy=(2π)n|Σ|.

What's annoying about this integral is that it contains the cross terms between vector components of y such as aijyiyj where aij=(Σ1)ij. If there were no cross terms, then the only terms are of the form aiiyi2, and we can split the multivariate integral into the product of univariate integrals. We can do this by diagonalizing Σ1.

Since Σ is real symmetric, we can diagonalize it by some orthogonal matrix U:

Σ=USU

where S=diag(s1,s2,,sn) is the diagonal matrix of the eigenvalues of Σ. Since Σ is positive definite, all the eigenvalues are positive. Accordingly (exercise!), the inverse matrix of Σ is diagonalized as

Σ1=US1U

where S1=diag(1/s1,1/s2,,1/sn) is the inverse matrix of S. Let's substitute this into the exponent of the density. We have

yΣ1y=yUS1Uy=(Uy)S1(Uy)=zS1z

where we set z=Uy. Since U is orthogonal, |U|=detU=±1. Therefore, dy=|detU|dz=dz, so that

Rnexp(12yΣ1y)dy=Rnexp(12zS1z)dz.

Since S1 is a diagonal matrix, the exponent becomes 

zS1z=z12/s1+z22/s2+zn2/sn.

Thus,

Rnexp(12zS1z)dz=Rni=1nexp(zi22si)dz=i=1nexp(zi22si)dzi=i=1n2πsi=(2π)ns1s2sn=(2π)n|Σ|

as

|Σ|=|USU|=|U||S||U|=|S|=s1s2sn.

Thus, (Eq:Goal) holds.

Comments

Popular posts from this blog

Birth process

Branching processes: Mean and variance

Informal introduction to formal logic