Reading for Today's Lecture:
Goals of Today's Lecture:
Review of end of last time
I defined independent events and then independent random variables:
Definition: Events Ai,
are
independent if
Definition: Random variables
are
independent if
New material
Proof:
Def'n:
P(A|B) = P(AB)/P(B) provided
.
Def'n: For discrete random variables X and Y the conditional
probability mass function of Y given X is
For absolutely continuous X the problem is that
P(X=x) = 0 for all
x so how can we define P(A| X=x) or
fY|X(y|x)?
The solution is to take a limit
Def'n:
iff
Def'n:
iff
(a column vector for later use) with the
Zi independent and each
.
In this case according to our theorem
where the superscript t denotes matrix transpose.
Def'n:
has a multivariate normal distribution if it
has the same distribution as
for some
,
some
matrix of constants A and
.
If the matrix A is singular then X will not have a density. If A is
invertible then we can derive the multivariate normal density
by the change of variables formula:
For which vectors
and matrices
is this a density?
Any
will work but if
then
where y=At x. The inequality is strict except for y=0 which
is equivalent to x=0. Thus
is a positive definite symmetric
matrix. Conversely, if
is a positive definite symmetric matrix
then there is a square invertible matrix A such that
so that
there is a
distribution. (The matrix A can be found
via the Cholesky decomposition, for example.)
More generally we say that X has a multivariate normal distribution
if it has the same distribution as
where now we remove the
restriction that A be non-singular. When A is singular X will not
have a density because there will exist an a such that
,
that is, so that X is confined to a hyperplane. It is still
true, however, that the distribution of X depends only on the matrix
in the sense that if
AAt = BBt then
and
have the same distribution.
Properties of the MVN distribution
1: All margins are multivariate normal: if
2: All conditionals are normal: the conditional distribution of X1
given X2=x2 is
3:
.
That is, an affine
transformation of a Multivariate Normal is normal.
Proof: Let
.
Then
are
independent N(0,1) so
is multivariate
standard normal. Note that
and
Thus
Thus we need only give a proof in the special case
and
.
Step 1: Define
Now it is easy to solve for Z from Y because
Zi = n-1/2Y1+Yi+1
for
.
To get Zn we use the identity
Now we use the change of variables formula to compute the density
of Y. It is helpful to let
denote the n-1 vector whose
entries are
.
Note that
Notice that this is a factorization into a function of y1 times a
function of
.
Thus
is independent of
.
Since sZ2 is a function of
we see that
and
sZ2 are independent.
Furthermore the density of Y1 is a multiple of the function of y1 in
the factorization above. But the factor in question is the standard normal
density so
.
We have now done the first 2 parts of the theorem. The third part is
a homework exercise but I will outline the derivation of the
density.
Suppose that
are independent N(0,1). We define the
distribution to be that of
.
Define angles
by
(These are spherical co-ordinates in n dimensions. The
values run
from 0 to
except for the last
whose values run from 0 to
.)
Then note the following derivative formulas
Finally the fourth part of the theorem is a consequence of the
first 3 parts of the theorem and the definition of the
distribution, namely, that
if it has the same distribution
as
However, I now derive the density of T in this definition:
I can differentiate this with respect to t by
simply differentiating the inner integral: