next_inactive up previous
Up: Stat 5102

Stat 5102 (Geyer) Final Exam

Problem 1

Looking at the simplest moment first,

$\displaystyle E(X_i) = \frac{\alpha}{\lambda} = \frac{\theta}{\theta} = 1,
$

we see that it is not a function of the unknown parameter and hence useless for finding a method of moments estimator. Moving up to the second central moment,

$\displaystyle \var(X_i) = \frac{\alpha}{\lambda^2} = \frac{\theta}{\theta^2}
= \frac{1}{\theta}
$

is a simple function of $ \theta$ and gives the method of moments estimator

$\displaystyle \hat{\theta}_n = \frac{1}{V_n}.
$

(or the same with $ V_n$ replaced by $ S^2_n$).

Problem 2

The formula for $ f$ should be familiar. It defines the location-scale family with base density $ g$ (Sections 4.1 and 9.2 of the course notes). The variables

$\displaystyle Y_i = \frac{X_i - \mu}{\sigma}
$

are i. i. d. with density $ g$. We get the same answer for ARE whether we compare $ \widetilde{X}_n$ and $ X{\mkern -13.5 mu}\overline{\phantom{\text{X}}}_n$ as estimators of $ \mu$ or whether we compare $ \widetilde{Y}_n$ and $ Y{\mkern -14.0 mu}\overline{\phantom{\text{Y}}}_n$ as estimators of zero.

$ \widetilde{Y}_n$ is asymptotically normal with variance

$\displaystyle \frac{1}{4 n g(0)^2} = \frac{4}{n}
$

(Corollary 7.28 in the notes), and $ Y{\mkern -14.0 mu}\overline{\phantom{\text{Y}}}_n$ is asymptotically normal with variance

$\displaystyle \frac{\var(Y)}{n} = \frac{\pi^2}{3 n}
$

Thus the ARE is either $ \pi^2 / 12$ or $ 12 / \pi^2$ depending on which way you form the ratio. The important point is that $ X{\mkern -13.5 mu}\overline{\phantom{\text{X}}}_n$ is the better estimator since

$\displaystyle \frac{\pi^2}{3} = 3.2899
$

is less than 4.

Alternate Solution

Things are only a little different if we don't realize we can give the answer for $ Y{\mkern -14.0 mu}\overline{\phantom{\text{Y}}}_n$ and $ \widetilde{Y}_n$ instead of $ X{\mkern -13.5 mu}\overline{\phantom{\text{X}}}_n$ and $ \widetilde{X}_n$.

Since $ X_i = \mu + \sigma Y_i$,

$\displaystyle E(X_i) = \mu
$

and

$\displaystyle \var(X_i) = \sigma^2 \var(Y_i) = \frac{\pi^2 \sigma^2}{3}
$

so

$\displaystyle X{\mkern -13.5 mu}\overline{\phantom{\text{X}}}_n \approx \NormalDis\left(\mu, \frac{\pi^2 \sigma^2}{3 n}\right)
$

Note that $ g$ is symmetric about zero but $ f$ is symmetric about $ \mu$, so $ \mu$ is both the population mean and the population median. And the asymptotic variance of $ \widetilde{X}_n$ is

$\displaystyle \frac{1}{4 n f(\mu)^2} = \frac{4 \sigma^2}{n}
$

and

$\displaystyle \widetilde{X}_n \approx \NormalDis\left(\mu, \frac{4 \sigma^2}{n}\right)
$

The ratio of asymptotic variances is the same as before.

Problem 3

We need to apply the delta method to the estimator $ g(X{\mkern -13.5 mu}\overline{\phantom{\text{X}}}_n)$, where

$\displaystyle g(x) = \frac{2 x}{1 - x}
$

has derivative

$\displaystyle g'(x) = \frac{2}{1 - x} + \frac{2 x}{(1 - x)^2}
= \frac{2}{(1 - x)^2}
$

Because $ g(X{\mkern -13.5 mu}\overline{\phantom{\text{X}}}_n)$ is a method of moments estimator it must have asymptotic mean $ \theta$ (you can check this if you like, but it is already part of the question statement and thus not a required calculation in the answer).

The asymptotic variance is

\begin{displaymath}
\begin{split}
g'(\mu)^2 \sigma^2
& =
g'\left(\frac{\theta...
...
& =
\frac{\theta (2 + \theta)^2}{2 (3 + \theta)}
\end{split}\end{displaymath}

Problem 4

If anyone is wondering whether sample size one is ``large,'' recall that a single Poisson random variable is approximately normal if the mean is large (Section F.3 in the appendices of the notes).

The obvious point estimate of $ \mu_Y - \mu_X$ is

$\displaystyle \hat{\mu}_Y - \hat{\mu}_X = Y - X = 410 - 320 = 90
$

which has variance

$\displaystyle \var(\hat{\mu}_Y - \hat{\mu}_X) = \var(Y) + \var(X) = 410 + 320 = 730
$

and standard deviation

$\displaystyle \sd(\hat{\mu}_Y - \hat{\mu}_X) = \sqrt{730} = 27.02
$

Thus the large sample confidence interval is

$\displaystyle 90 \pm 1.96 \times 27.02
$

or

$\displaystyle 90 \pm 52.96
$

or

$\displaystyle (37.0, 143.0)
$

Problem 5

The density of the data is

$\displaystyle f(x \mid \theta)
=
\frac{\Gamma(\theta + 1)}{\Gamma(\theta) \Gamma(1)} x^{\theta - 1}
=
\theta x^{\theta - 1},
\qquad 0 < x < 1
$

The likelihood for a sample of size $ n$ is

$\displaystyle L_n(\theta)
=
\prod_{i = 1}^n \theta x_i^{\theta - 1}
=
\theta^n ...
...theta - 1}
=
\theta^n \exp\left( (\theta - 1) \sum_{i = 1}^n \log(x_i) \right)
$

or, if we introduce the variables $ y_i = \log(x_i)$,

$\displaystyle L_n(\theta)
=
\theta^n e^{(\theta - 1) n \bar{y}_n}
=
\theta^n e^{\theta n \bar{y}_n} e^{- n \bar{y}_n}
$

The last term can be dropped, since it does not contain the parameter, giving

$\displaystyle L_n(\theta)
=
\theta^n e^{\theta n \bar{y}_n}
$

The log likelihood is thus

$\displaystyle l_n(\theta)
=
n \log(\theta) + n \theta \bar{y}_n
$

which has derivatives

\begin{displaymath}
\begin{split}
l_n'(\theta)
& =
\frac{n}{\theta} + n \bar{y}_n
\\
l_n''(\theta)
& =
- \frac{n}{\theta^2}
\end{split}\end{displaymath}

Since the latter does not depend on the data, it is the same as it's expectation, so observed and expected Fisher information are the same

$\displaystyle J_n(\theta) = I_n(\theta) = \frac{n}{\theta^2}
$

Problem 6

The likelihood is

$\displaystyle L_n(p)
=
\prod_{i = 1}^n p (1 - p)^{x_i}
=
p^n (1 - p)^{\sum_i x_i}
$

Note that this is the same functional form as a binomial likelihood. The only difference is the statistics. Here there are $ n$ successes and $ \sum_i x_i$ failures. Since Bayesian inference depends only on the log likelihood, there is no difference between the calculations here an in Example 11.2.3 in the notes. The posterior distribution is $ \BetaDis(n + 1, 1 + \sum_i x_i)$ and the posterior mean is

$\displaystyle E(p \mid x_1, \ldots, x_n)
=
\frac{n + 1}{n + 2 + \sum_i x_i}
$

Problem 7

(a)

The models in question have polynomials of degree 1, 2, 3, 4, and 5 as regression functions.

To be more precise, the regression functions are of the form

$\displaystyle g(x) = \beta_0 + \sum_{i = 1}^k \beta_i x^i
$

for $ i = 1$, 2, 3, 4, 5. Strictly speaking, the model assumptions include the assumption of i. i. d. normal errors, but almost no one said this and no deduction was made for not saying it.

(b)

Because linear functions are special cases of quadratic, and so forth. You obtain the models of lower degree by setting the coefficients of higher powers of $ x$ to zero in the larger models.

(c)

Starting at the bottom of the ANOVA table and reading up

Thus we conclude that the cubic model (Model 3) is correct, which means its supermodels (Model 4 and Model 5) must also be correct. Or to be more finicky we conclude that these data do not give any evidence that these models are incorrect. And we conclude that the quadratic model (Model 2) and its submodel (Model 1) are incorrect. The evidence for that latter conclusion is very strong (repeating what was said above, $ P < 2 \times 10^{-16}$).

Many people were confused by ``correct'' and ``incorrect.'' If a model is correct, then so is every supermodel. If a model is incorrect, then so is every submodel. Hence in a nested sequence of models, there is a smallest correct model (here model 3) and all the models above it are also correct, but all the models below it are incorrect.

Problem 8

\begin{displaymath}
\begin{split}
f(y \mid p)
& =
p (1 - p)^{y - 1}
\\
& ...
...1 - p)]
\\
& =
\exp[y \log(1 - p) + \logit(p)]
\end{split}\end{displaymath}

where $ \logit(p)$ is defined by equation (12.77a) in the notes.

This clearly fits the form of equation (12.78) in the notes with

\begin{displaymath}
\begin{split}
\theta & = \log(1 - p)
\\
\phi & = 1
\\
w & = 1
\\
b(\theta) & = - \logit(p)
\end{split}\end{displaymath}

Of course, the last equation doesn't by itself define $ b(\theta)$. To do that we need to know $ p$ as a function of $ \theta$, that is, we have to solve the first equation above for $ p$ giving

$\displaystyle p = 1 - e^\theta
$

and plugging that in to the equation for $ b(\theta)$ giving

\begin{displaymath}
\begin{split}
b(\theta)
& =
- \logit(p)
\\
& =
- \lo...
...eta}\right)
\\
& =
\theta - \log(1 - e^\theta)
\end{split}\end{displaymath}

or, if you prefer, starting just above the last line

\begin{displaymath}
\begin{split}
\hphantom{b(\theta)}
& =
\log\left(\frac{1}...
...a} - 1}\right)
\\
& =
- \log(e^{- \theta} - 1)
\end{split}\end{displaymath}

Additional Stuff

This is just added for my curiosity and perhaps to go in the homework problems some future semester.

\begin{displaymath}
\begin{split}
b'(\theta)
& =
\frac{e^{- \theta}}{e^{- \th...
...& =
\frac{1}{1 - e^\theta}
\\
& =
\frac{1}{p}
\end{split}\end{displaymath}

which is indeed the mean of $ Y$ given in Section B.1.8 of the appendices to the notes. Furthermore

\begin{displaymath}
\begin{split}
b''(\theta)
& =
\frac{e^\theta}{(1 - e^\theta)^2}
\\
& =
\frac{1 - p}{p^2}
\end{split}\end{displaymath}

which is indeed the variance of $ Y$ given in Section B.1.8 of the appendices to the notes. So the GLM theory works and Lemma 12.20 in the notes is true (as, of course, it must be since it is proved in the notes).


next_inactive up previous
Up: Stat 5102
Charles Geyer 2001-05-10