conditional distribution of multi-variate normal variable
Theorem.
Let be a random variable, taking values in , normally distributed with a non-singular covariance matrix and a mean of zero.
Suppose is defined by for some linear transformation of maximum rank. ( to denotes the transpose operator.)
Then the distribution of conditioned on is multi-variate normal, with conditional means and covariances of:
If , so that is simply a vector in , these formulas reduce to:
If does not have zero mean, then the formula for is modified by adding and replacing by , and the formula for is unchanged.
Proof.
We split up into two stochastically independent parts, the first part containing exactly the information embodied in . Then the conditional distribution of given is simply the unconditional distribution of the second part that is independent of .
To this end, we first change variables to express everything in terms of a standard multi-variate normal . Let be a βsquare rootβ factorization of the covariance matrix , so that:
We let be the orthogonal projection onto the range of , and decompose into orthogonal components:
It is intuitively obvious that orthogonality of the two random normal vectors implies their stochastic independence. To show this formally, observe that the Gaussian density function for factors into a product:
We can construct an orthonormal system of coordinates on under which the components for are completely disjoint from those components of . On the other hand, the densities for , , and remain invariant even after changing coordinates, because they are radially symmetric. Hence the variables and are separable in their joint density and they are independent.
embodies the information in the linear combination . For we have the identity:
The last term is null because is orthogonal to the range of by definition. (Equivalently, lies in the kernel of .) Thus can always be recovered by a linear transformation on .
Conversely, completely determines , from the analytical expression for that we now give. In general, the orthogonal projection onto the range of an injective transformation is . Applying this to , we have
We see that .
We have proved that conditioning on and are equivalent, and so:
and
using the defining property of orthogonal projections.
Now we express the result in terms of , and remove the dependence on the transformation (which is not uniquely defined from the covariance matrix):
and
Of course, the conditional distribution of given is the same as that of , which is multi-variate normal.
Title | conditional distribution of multi-variate normal variable |
---|---|
Canonical name | ConditionalDistributionOfMultivariateNormalVariable |
Date of creation | 2013-03-22 18:39:09 |
Last modified on | 2013-03-22 18:39:09 |
Owner | stevecheng (10074) |
Last modified by | stevecheng (10074) |
Numerical id | 5 |
Author | stevecheng (10074) |
Entry type | Theorem |
Classification | msc 62E15 |
Classification | msc 60E05 |