We extend the definition of the Normal Distribution to higher dimensions. Indeed
where is a positive symmetric definite matrix.
Positive Symmetric Definite
A matrix is symmetric if and positive definite if for every nonzero real column vector , .
https://en.wikipedia.org/wiki/Definite_matrix
- For any real, invertible matrix , the product is PSD. This is called the Cholesky Factorization.
Bivariate Normal Distribution
Special case when . We say that has bivariate normal distribution if the joint density
Example 1
If and are independent then have bivariate normal distribution:
This is because we have the Joint Distribution: in which the resulting exponent is three matrices with
Properties
Normalization
- Let and be independent.
- Let be a 2x2 PSD where .
Then
has bivariate normal distribution,
Conversely, if
then
which is how we can “normalize it”.
Proof:
We have that
which is just
Joint Conversion
What is the joint density of ? We can define the prior function as , where
where
where we multiply by the determinant because changes the volume or “space” in since we are switching from one coordinate system to another.
- The determinant encodes the notion of area and volume
- Multiplying by it fixes the transformation change so that the probabilities match in both coordinate systems.
Back to the proof, we can apply the inverse to the map. Let the map be .
- is an affine transformation (linear transformation given by ) then translated by .
- is the Jacobian operator on .
Upon changing variables and application of the density function,
We read from here that has bivariate normal distribution with
Variance and Covariance
Suppose
then
which is the mean vector and the covariance matrix
Proof
We use joint conversion and normalization . So, we have
where and and are independent. We denote . Then
Then to find the variances:
and likewise for . Next,
As , we get
which follows from what we calculated before. For the means, we have that
which gives us:
Equivalent Characterization of Bivariate Normal
have a bivariate normal distribution if and only if , is a normal random variable.
Proof:
Forward Direction. Upon conversion to matrices through joint conversion:
where . This gives us . But this is normal for any by convolutions which is shown in Sum of 2 Independent RVs.
Reverse Direction. Better explained with the Fourier transformation and is not covered.
Conditional Distribution
Let . Then the conditional distribution of given is
where is the Correlation.
Corollary: Conditional Expectation
Proof:
If has bivariate normal distribution, then the Joint Distribution is determined by
In particular, if then are independent.
In general, think of as an inner product for the bivariate normal distribution. So,
We write such that . Then
then
since
is bivariate normal, we deduce that is independent of . So,
and has normal distribution and