Problems on estimation, sufficiency, Fisher information, and hypothesis testing
Instructions
Let be i.i.d. from where θ > 0 is unknown.
(1) Find a sufficient statistic for θ.
(2) Show that is sufficient for θ.
(3) Is sufficient for θ?
Look for factorization of the likelihood so the data only enters through a low-dimensional statistic that captures all parameter information.
Let be i.i.d. from a distribution with density:
(1) Find the MLE of θ.
(2) Is the MLE unbiased? If not, find an unbiased estimator based on the MLE.
(3) Find the variance of the MLE.
Write the likelihood, take logs to simplify differentiation, and solve the score equation while checking parameter constraints.
Let be i.i.d. Poisson(λ).
(1) Find the Fisher information for a single observation.
(2) Find the Cramér-Rao lower bound for unbiased estimators of λ.
(3) Show that attains the CRLB.
(4) Find the CRLB for estimating .
Compute Fisher information from the score or Hessian, then compare estimator variance to the Cramer-Rao lower bound.
Let be i.i.d. from Gamma(α, β) with density:
Find the method of moments estimators for α and β.
Match sample moments to theoretical moments and solve the resulting system for the unknown parameter values.
Let be i.i.d. N(μ, σ²) where σ² is known.
Test vs .
(1) Derive the likelihood ratio test statistic.
(2) Find the rejection region for significance level α.
(3) Show the LRT is equivalent to the z-test.
Build the unrestricted and restricted likelihoods, form their ratio, and reject when the data looks too unlikely under the null model.
Let be i.i.d. from Exponential(λ).
(1) Find a pivot based on .
(2) Construct a 95% confidence interval for λ.
(3) Find the expected length of this confidence interval.
Choose the right pivot or sampling distribution, then invert the probability statement to produce lower and upper bounds.
Let be i.i.d. Bernoulli(p).
(1) Show that is complete and sufficient.
(2) Find the UMVUE of p.
(3) Find the UMVUE of p(1-p).
Combine sufficiency, completeness, and unbiasedness so the final estimator is justified by Lehmann-Scheffe rather than guesswork.
Let be i.i.d. Poisson(λ). Assume a Gamma(α, β) prior for λ.
(1) Find the posterior distribution of λ.
(2) Find the Bayes estimator under squared error loss.
(3) Find a 95% Bayesian credible interval for λ.
Start with the prior and likelihood, form the posterior, and then read off the Bayes estimator under the stated loss function.
Review Common Distribution Families
Revisit the theory that supports these worked practice problems.
Review Point Estimation and Cramer-Rao Bound
Revisit the theory that supports these worked practice problems.
Review Sufficient and Complete Statistics
Revisit the theory that supports these worked practice problems.
Review Bayesian Statistics and Inference
Revisit the theory that supports these worked practice problems.