Idea of MLE: “likelihood” is the probability of the dataset appears, so the parameter that maximizes the likelihood might be the true parameter.
The Definition of MLE
Step 1, make assumptions:
- assume the distribution is …, and the PDF/PMF is
Now estimate the parameters by maximizing the likelihood.
Likelihood: the product of the possibility / possibility density of the data / dataset.
Given a sample dataset , the likelihood is a function of :
Since log is monotone, sometimes we use log-likelihood:
Find the that maximize the possibility of occurs:
The property of MLE
consistency of MLE
the optimized is converge in probability to the true parameter
asymptotic normality of MLE
估计误差的分布会趋近于正态分布
asymptotic efficiency of MLE
Invariance of MLE
The MLE of is
Define the likelihood of as the supremum of with :
And the optimized should be the one that makes be supremum of , which is the supremum of :
and the likelihood of is:
which means , so can be