Method of maximum likelihood
Finds a value for such that it gives the maximum probability of observing the observed data in comparison to other values of
If are observed values of a random sample from a population with the parameter , the likelihood function of is
The maximum likelihood estimate (MLE) of is the value of that maximizes the likelihood function
Under the regular case, we use the log-likelihood function, as we will only need to differentiate a sum of functions instead of a product
And by a lemma, the that maximizes also maximizes
Properties
- MLE of is a sufficient statistic, if one exists, then MLE is a function of it
- is known to be asymptotically efficient
- Invariance principle: if is the MLE of , then is the MLE of the function
- Lack of uniqueness: there could be more than one MLE
Example
An experiment with 6 coin tosses, 2 heads
General pdf with arbitrary parameter:
Different values of give us different probabilities of getting that sample
Finding the MLE of
On a irregular case,
Case: 2+ parameters (with hessian)
MLE of
You can't use 'macro parameter character #' in math mode\text{if} \begin{bmatrix} \frac{\partial^2 l }{\partial \mu^2} & \frac{\partial { #2l} }{\partial \mu \partial \sigma^2} \\ \frac{\partial^2 l}{\partial \sigma^2 \partial \mu} & \frac{\partial { #2} l}{\partial (\sigma^2 )^2} \end{bmatrix} < 0, \text{ then our} \hat{\mu}, \hat{\sigma}^2 \text{ are MLEs of }\mu, \sigma^2 Then the MLE of
And the MLE of
Case: area of solutions