We extend the definition of the Normal Distribution to higher dimensions. Indeed

where is a positive symmetric definite matrix.

Positive Symmetric Definite

A matrix is symmetric if and positive definite if for every nonzero real column vector , .
https://en.wikipedia.org/wiki/Definite_matrix

  • For any real, invertible matrix , the product is PSD. This is called the Cholesky Factorization.

Bivariate Normal Distribution

Special case when . We say that has bivariate normal distribution if the joint density

Example 1

If and are independent then have bivariate normal distribution:

This is because we have the Joint Distribution: in which the resulting exponent is three matrices with

Properties

Normalization

  • Let and be independent.
  • Let be a 2x2 PSD where .

Then

has bivariate normal distribution,

Conversely, if

then

which is how we can “normalize it”.

Proof:

We have that

which is just


Joint Conversion

What is the joint density of ? We can define the prior function as , where

where

where we multiply by the determinant because changes the volume or “space” in since we are switching from one coordinate system to another.

  • The determinant encodes the notion of area and volume
  • Multiplying by it fixes the transformation change so that the probabilities match in both coordinate systems.

Back to the proof, we can apply the inverse to the map. Let the map be .

  • is an affine transformation (linear transformation given by ) then translated by .
  • is the Jacobian operator on .

Upon changing variables and application of the density function,

We read from here that has bivariate normal distribution with

Variance and Covariance

Suppose

then

which is the mean vector and the covariance matrix

Proof

We use joint conversion and normalization . So, we have

where and and are independent. We denote . Then

Then to find the variances:

and likewise for . Next,

As , we get

which follows from what we calculated before. For the means, we have that

which gives us:

Equivalent Characterization of Bivariate Normal

have a bivariate normal distribution if and only if , is a normal random variable.

Proof:

Forward Direction. Upon conversion to matrices through joint conversion:

where . This gives us . But this is normal for any by convolutions which is shown in Sum of 2 Independent RVs.

Reverse Direction. Better explained with the Fourier transformation and is not covered.

Conditional Distribution

Let . Then the conditional distribution of given is

where is the Correlation.

Corollary: Conditional Expectation

Proof:

If has bivariate normal distribution, then the Joint Distribution is determined by

In particular, if then are independent.

In general, think of as an inner product for the bivariate normal distribution. So,

We write such that . Then

then

since

is bivariate normal, we deduce that is independent of . So,

and has normal distribution and