next up previous


Postscript version of this file

STAT 801 Lecture 15

Reading for Today's Lecture:

Goals of Today's Lecture:

Today's notes

Finding (good) preliminary Point Estimates

Method of Moments

Basic strategy: set sample moments equal to population moments and solve for the parameters.

Gamma Example

The Gamma( $\alpha,\beta$) density is

\begin{displaymath}f(x;\alpha,\beta) =
\frac{1}{\beta\Gamma(\alpha)}\left(\frac{...
...ta}\right)^{\alpha-1}
\exp\left[-\frac{x}{\beta}\right] 1(x>0)
\end{displaymath}

and has

\begin{displaymath}\mu_1 = \alpha\beta
\end{displaymath}

and

\begin{displaymath}\mu_2^\prime = \alpha(\alpha+1)\beta^2.\end{displaymath}

This gives the equations
\begin{align*}\alpha\beta & = \overline{X}
\\
\alpha(\alpha+1)\beta^2 & = \overline{X^2}
\end{align*}
or
\begin{align*}\alpha\beta & = \overline{X}
\\
\alpha\beta^2 & = \overline{X^2} - \overline{X}^2
\end{align*}
Divide the second equation by the first to find the method of moments estimate of $\beta$ is

\begin{displaymath}\tilde\beta = (\overline{X^2} - \overline{X}^2)/\overline{X}
\end{displaymath}

Then from the first equation get

\begin{displaymath}\tilde\alpha = \overline{X}/\tilde\beta= (\overline{X})^2/(\overline{X^2} - \overline{X}^2)
\end{displaymath}

These equations are much easier to solve than the likelihood equations. The latter involve the function

\begin{displaymath}\psi(\alpha) = \frac{d}{d\alpha} \log(\Gamma(\alpha))
\end{displaymath}

called the digamma function. The score function in this problem has components

\begin{displaymath}U_\beta = \frac{\sum X_i}{\beta^2} - n \alpha / \beta
\end{displaymath}

and

\begin{displaymath}U_\alpha = -n\psi(\alpha) +\sum\log(X_i) -n\log(\beta)
\end{displaymath}

You can solve for $\beta$ in terms of $\alpha$ to leave you trying to find a root of the equation

\begin{displaymath}-n\psi(\alpha) +\sum\log(X_i) -n \log(\sum X_i/(n\alpha)) = 0
\end{displaymath}

To use Newton Raphson on this you begin with the preliminary estimate $\hat\alpha_1 = \tilde\alpha$ and then compute iteratively

\begin{displaymath}\hat\alpha_{k+1} = \frac{\overline{\log(X)} - \psi(\hat\alpha...
...overline{ X}/\hat\alpha_k}{1/\alpha-\psi^\prime(\hat\alpha_k)}
\end{displaymath}

until the sequence converges. Computation of $\psi^\prime$, the trigamma function, requires special software. Web sites like netlib and statlib are good sources for this sort of thing.

Optimality theory for point estimates

Why bother doing the Newton Raphson steps? Why not just use the method of moments estimates? The answer is that the method of moments estimates are not usually as close to the right answer as the mles.

Rough principle: A good estimate $\hat\theta$ of $\theta$ is usually close to $\theta_0$ if $\theta_0$ is the true value of $\theta$. Closer estimates, more often, are better estimates.

This principle must be quantified if we are to ``prove'' that the mle is a good estimate. In the Neyman Pearson spirit we measure average closeness.

Definition: The Mean Squared Error (MSE) of an estimator $\hat\theta$ is the function

\begin{displaymath}MSE(\theta) = E_\theta[(\hat\theta-\theta)^2]
\end{displaymath}

Standard identity:

\begin{displaymath}MSE = {\rm Var}_\theta(\hat\theta) + Bias_{\hat\theta}^2(\theta)
\end{displaymath}

where the bias is defined as

\begin{displaymath}Bias_{\hat\theta}(\theta) = E_\theta(\hat\theta) - \theta \, .
\end{displaymath}

Primitive example: I take a coin from my pocket and toss it 6 times. I get HTHTTT. The MLE of the probability of heads is

\begin{displaymath}\hat{p} = X/n
\end{displaymath}

where X is the number of heads. In this case I get $\hat{p}
=\frac{1}{3}$.

An alternative estimate is $\tilde p = \frac{1}{2}$. That is, $\tilde p$ ignores the data and guesses the coin is fair. The MSEs of these two estimators are

\begin{displaymath}MSE_{\text{MLE}} = \frac{p(1-p)}{6}
\end{displaymath}

and

MSE0.5 = (p-0.5)2

If p is between 0.311 and 0.689 then the second MSE is smaller than the first. For this reason I would recommend use of $\tilde p$ for sample sizes this small.

Now suppose I did the same experiment with a thumbtack. The tack can land point up (U) or tipped over (O). If I get UOUOOO how should I estimate p the probability of U? The mathematics is identical to the above but it seems clear that there is less reason to think $\tilde p$ is better than $\hat p$ since there is less reaon to believe $0.311 \le p \le 0.689$ than with a coin.

Unbiased Estimation

The problem above illustrates a general phenomenon. An estimator can be good for sme values of $\theta$ and bad for others. When comparing $\hat\theta$ and $\tilde\theta$, two estimators of $\theta$ we will say that $\hat\theta$ is better than $\tilde\theta$ if it has uniformly smaller MSE:

\begin{displaymath}MSE_{\hat\theta}(\theta) \le MSE_{\tilde\theta}(\theta)
\end{displaymath}

for all $\theta$. Normally we also require that the inequality be strict for at least one $\theta$.

The definition raises the question of the existence of a best estimate - one which is better than every other estimator. There is no such estimate. Suppose $\hat\theta$ were such a best estimate. Fix a $\theta^*$ in $\Theta$ and let $\tilde p\equiv \theta^*$. Then the MSE of $\tilde p$ is 0 when $\theta=\theta^*$. Since $\hat\theta$ is better than $\tilde p$ we must have

\begin{displaymath}MSE_{\hat\theta}(\theta^*) = 0
\end{displaymath}

so that $\hat\theta=\theta^*$ with probability equal to 1. This makes $\hat\theta=\tilde\theta$. If there are actually two different possible values of $\theta$ this gives a contradiction; so no such $\hat\theta$ exists.

Principle of Unbiasedness: A good estimate is unbiased, that is,

\begin{displaymath}E_\theta(\hat\theta) \equiv = 0 \, .
\end{displaymath}

WARNING: In my view the Principle of Unbaisedness is a load of hog wash.

For an unbiased estimate the MSE is just the variance.

Definition: An estimator $\hat\phi$ of a parameter $\phi=\phi(\theta)$ is Uniformly Minimum Variance Unbiased (UMVU) if, whenever $\tilde\phi$ is an unbiased estimate of $\phi$ we have

\begin{displaymath}{\rm Var}_\theta(\hat\phi) \le {\rm Var}_\theta(\tilde\phi)
\end{displaymath}

We call $\hat\phi$ the UMVUE. (`E' is for Estimator.)

The point of having $\phi(\theta)$ is to study problems like estimating $\mu$ when you have two parameters like $\mu$ and $\sigma$ for example.

Cramér Rao Inequality

If $\phi(\theta)=\theta$ we can derive some informtion from the identity

\begin{displaymath}E_\theta(T) \equiv\theta
\end{displaymath}

When we worked with the score function we derived some information from the identity

\begin{displaymath}\int f(x,\theta) dx \equiv 1
\end{displaymath}

by differentiation and we do the same here. If T=T(X) is some function of the data X which is unbiased for $\theta$ then

\begin{displaymath}E_\theta(T) = \int T(x) f(x,\theta) dx \equiv \theta
\end{displaymath}

Differentiate both sides to get
\begin{align*}1 & = \frac{d}{d\theta} \int T(x) f(x,\theta) dx
\\
&= \int T(x)...
... \log(f(x,\theta))
f(x,\theta) dx
\\
& = E_\theta( T(X) U(\theta))
\end{align*}
where U is the score function. Since we already know that the score has mean 0 we see that

\begin{displaymath}{\rm Cov}_\theta(T(X),U(\theta)) = 1
\end{displaymath}

Now remember that correlations are between -1 and 1 or

\begin{displaymath}1=\vert{\rm Cov}_\theta(T(X),U(\theta))\vert \le \sqrt{{\rm Var}_\theta(T) {\rm
Var}_\theta(U(\theta))}
\end{displaymath}

Squaring gives the inequality

\begin{displaymath}{\rm Var}_\theta(T) \ge \frac{1}{I(\theta)}
\end{displaymath}

which is called the Cramér Rao Lower Bound. The inequality is strict unless the correlation is 1 which would require that

\begin{displaymath}U(\theta) = A(\theta) T(X) + B(\theta)
\end{displaymath}

for some non-random constants A and B (which might depend on $\theta$.) This would prove that

\begin{displaymath}\ell(\theta) = A^*(\theta) T(X) + B^*(\theta) + C(X)
\end{displaymath}

for some further constants A* and B* and finally

\begin{displaymath}f(x,\theta) = h(x) e^{A*(\theta)T(x)+B^*(\theta)}
\end{displaymath}

for h=eC.

Summary of Implications

What can we do to find UMVUEs when the CRLB is a strict inequality?

Example: Suppose X has a Binomial(n,p) distribution. The score function is

\begin{displaymath}U(p) =
\frac{1}{p(1-p)} X - \frac{n}{1-p}
\end{displaymath}

Thus the CRLB will be strict unless T=cX for some c. If we are trying to estimate p then choosing c=n-1 does give an unbiased estimate $\hat p = X/n$ and T=X/n achieves the CRLB so it is UMVU.

A different tactic proceeds as follows. Suppose T(X) is some unbiased function of X. Then we have

\begin{displaymath}E_p(T(X)-X/n) \equiv 0\end{displaymath}

because $\hat p = X/n$ is also unbiased. If h(k) = T(k)-k/n then

\begin{displaymath}E_p(h(X)) = \sum_{k=0}^n h(k) \left(\begin{array}{c} n \\ k \end{array}\right) p^k (1-p)^{n-k}
\equiv 0
\end{displaymath}

The left hand side of the $\equiv$ sign is a polynomial function of p as is the right. Thus if the left hand side is expanded out the coefficient of each power pk is 0. The constant term occurs only in the term k=0 and its coefficient is

\begin{displaymath}h(0) \left(\begin{array}{c} n \\ 0 \end{array} \right) = h(0)
\end{displaymath}

Thus h(0) = 0. Now p1=p occurs only in the term k=1 with coefficient nh(1) so h(1)=0. Since the terms with k=0 or 1 are 0 the quantity p2 occurs only in the term with k=2 with coefficient

n(n-1)h(2)/2

so h(2)=0. We can continue in this way to see that in fact h(k)=0 for each k and so the only unbiased function of X is X/n.

Now a Binomial random variable is just of sum of n iid Bernouilli(p) random variables. If $Y_1,\ldots,Y_n$ are iid Bernouilli(p) then $X=\sum Y_i$ is Binomial(n,p). Could we do better by than $\hat p = X/n$ by trying $T(Y_1,\ldots,Y_n)$ for some other function T?

Let's consider the case n=2 so that there are 4 possible values for Y1,Y2. If h(Y1,Y2) = T(Y1,Y2) - [Y1+Y2]/2 then again

\begin{displaymath}E_p(h(Y_1,Y_2)) \equiv 0
\end{displaymath}

and we have

Ep( h(Y1,Y2)) = h(0,0)(1-p)2 + [h(1,0)+h(0,1)]p(1-p)+ h(1,1) p2

This can be rewritten in the form

\begin{displaymath}\sum_{k=0}^n w(k) \left(\begin{array}{c} n \\ k \end{array}\right)
p^k(1-p)^{n-k}
\end{displaymath}

where w(0)=h(0,0), w(1) =[h(1,0)+h(0,1)]/2 and w(2) = h(1,1). Just as before it follows that w(0)=w(1)=w(2)=0. This argument can be used to prove that for any unbiased estimate $T(Y_1,\ldots,Y_n)$ we have that the average value of $T(y_1,\ldots,y_n)$ over vectors $y_1,\ldots,y_n$ which have exactly k 1s and n-k 0s is k/n. Now let's look at the variance of T:
\begin{align*}{\rm Var(T)} = & E_p( [T(Y_1,\ldots,Y_n) - p]^2)
\\
=& E_p( [T(...
...
& + 2E_p( [T(Y_1,\ldots,Y_n)
-X/n][X/n-p])
\\ & + E_p([X/n-p]^2)
\end{align*}
I claim that the cross product term is 0 which will prove that the variance of T is the variance of X/n plus a non-negative quantity (which will be positive unless $T(Y_1,\ldots,Y_n) \equiv X/n$). We can compute the cross product term by writing
\begin{multline*}E_p( [T(Y_1,\ldots,Y_n)-X/n][X/n-p])
\\ \sum_{y_1,\ldots,y_n}
...
...,y_n)-\sum y_i/n][\sum y_i/n -p] p^{\sum y_i} (1-p)^{n-\sum
y_i}
\end{multline*}
We can do the sum by summing over those $y_1,\ldots,y_n$ whose sum is an integer x and then summing over x. We get
\begin{multline*}E_p( [T(Y_1,\ldots,Y_n)-X/n][X/n-p])
\\
= \sum_{x=0}^n
\sum...
...m y_i=x} [T(y_1,\ldots,y_n)-x/n]\right][x/n -p]p^{x}
(1-p)^{n-x}
\end{multline*}

We have already shown that the sum in [] is 0!

This long, algebraically involved, method of proving that $\hat p = X/n$ is the UMVUE of p is one special case of a general tactic.

To get more insight I begin by rewriting
\begin{multline*}E_p(T(Y_1,\ldots,Y_n))
\\
\sum_{x=0}^n \sum_{\sum y_i = x} T...
...
\left(\begin{array}{c} n \\ x \end{array}\right) p^x(1-p)^{n-x}
\end{multline*}
Notice that the large fraction in this formula is the average value of T over values of y when $\sum y_i$ is held fixed at x. Notice that the weights in this average do not depend on p. Notice that this average is actually
\begin{multline*}E(T(Y_1,\ldots,Y_n\vert X=x))
\\
\sum_{y_1,\ldots,y_n} T(y_1,\ldots,y_n)
P(Y_1=y_1,\ldots,Y_n=y_n\vert X=x)
\end{multline*}
Notice that the conditional probabilities do not depend on p. In a sequence of Binomial trials if I tell you that 5 of 17 were heads and the rest tails the actual trial numbers of the 5 Heads are chosen at random from the 17 possibilities; all of the 17 choose 5 possibilities have the same chance and thsi chance does not depend on p.

Notice that in this problem with data $Y_1,\ldots,Y_n$ the log likelihood is

\begin{displaymath}\ell(p) = \sum Y_i \log(p) - (n-\sum Y_i) \log(1-p)
\end{displaymath}

and

\begin{displaymath}U(p) =
\frac{1}{p(1-p)} X - \frac{n}{1-p}
\end{displaymath}

as before. Again we see that the CRLB will be a strict inequality except for multiples of X. Since the only unbiased multiple of X is $\hat p = X/n$ we see again that $\hat p$ is UMVUE for p.


next up previous



Richard Lockhart
1998-10-29