ECON 626: Problem Set 4
\[ \def\R{{\mathbb{R}}} \def\Er{{\mathrm{E}}} \]
Problem 1
The exponential distribution has density (with respect to Lesbegue measure) \[ f(X;\lambda) = \frac{1}{\lambda} e^{-X/\lambda} 1\{X > 0\} \] for \(\lambda>0\). Suppose \(X_1, ... , X_n\) are independently exponential\((\lambda)\) distributed .
Show that the maximum likelihood estimator for \(\lambda\) is \(\hat{\lambda}^{MLE} = \frac{1}{n} \sum_{i=1}^n X_i\)
Derive the Cramér Rao lower bound for any unbiased estimator for \(\lambda\). Is \(\hat{\lambda}^{MLE}\) a minimum variance unbiased estimator?
Find the most powerful test of size \(\alpha\) for testing \(H_0: \lambda = \lambda_0\) versus \(H_1:\lambda = \lambda_1\).
Problem 2
Suppose \(X_1, ... , X_n\) are independently uniformly distributed on \((0,\theta)\).
Show that \(2 \bar{X} = \frac{2}{n} \sum_{i=1}^n X_i\) is an unbiased estimator for \(\theta\).
Show that \(\hat{\theta} = \frac{n+1}{n} \max_{1 \leq i \leq n} X_i\) is an unbiased estimator for \(\theta\) with \(Var(\hat{\theta}) < Var(2 \bar{X})\).
Is there a Cramér Rao lower bound for the variance of \(\theta\)? Why or why not?
(Optional and challenging) Show that \(\hat{\theta} = \frac{n+1}{n} \max_{1 \leq i \leq n} X_i\) is a minimum variance unbiased estimator for \(\theta\).