<< 6.1 Maximum Likelihood Estimation.md | 6.3 Maximum Likelihood Tests >>

Assumptions 6.2.1: Additional regularity conditions 1

  • : The pdf is twice differentiable as a function of .
  • : The integral can be differentiated twice under the integral sign as a function of .

Definition: Score function

Definition

Let

Then the score function is defined as $S(x;\theta) = \frac{\partial}{\partial \theta} \ln f(x;\theta)$$

Definition: Fisher information

Definition

Let

Then the Fisher information is defined as:

Theorem 6.2.1: Rao-Cramér lower bound

Definition

Let

Assume regularity conditions and additional regularity conditions 1 hold.

Then $ \operatorname{Var}(Y)\geq \boxed{\frac{[k’(\theta)]^2}{nI(\theta)}} $$

  • We say is the Rao-Cramer lower bound of

Corollary 6.2.1: Rao-Cramér bound for unbiased estimators

Under the same conditions as Theorem 6.2.1 Rao-Cramér lower bound

If is an unbiased estimator of (so that and )

Then

Important

We call the ratio of Rao-Cramer Lower Bound divided by the efficiency of : This means

  1. If efficiency = 1, then is efficient
  2. If efficiency approaches 1, then is asymptotically efficient

Thus is efficient if and only if Assumptions 6.2.1 Additional regularity conditions 1 holds

Definition 6.2.1: Efficient estimator

Definition

Let : Unbiased estimator of parameter

Then is an efficient estimator attains the Rao-Cramér lower bound

Assumptions 6.2.2: Additional regularity conditions 2

  • : The pdf is three times differentiable as a function of . Further, for all , there exist a constant and a function such that with , for all and all in the support of .

Theorem 6.2.2

Assume

  • : Random samples, with
    • pdf , for
    • All regularity conditions are satisfied.

If fisher information satisfies

Then any consistent sequence of solutions of the mle equations satisfies