[A, SfS] Chapter 4: Estimation: 4.2: Maximum Likelihood Estimation
Maximum Likelihood Estimation
Maximum Likelihood Estimation
In this lesson, you will learn about a method for finding an optimal estimator for an unknown parameter.
If we know that are the result of independent measurements of a random variable having a probability distribution that depends on some parameter with pmf or pdf , we would like to obtain some estimator of which depends on and has the smallest MSE possible.
The most popular method for finding such an estimator is the Method of Maximum Likelihood, and the estimator it produces is called the Maximum Likelihood Estimator, or MLE.
Method of Maximum Likelihood
The idea behind the Method of Maximum Likelihood is to find the value for the parameter that provides the best explanation for the data we observed.
Process of Maximum Likelihood Estimation
For simplicity, assume we have a continuous random variable with pdf depending on an unknown parameter , and are the result of independent observations of on a random sample of size . If is discrete with pmf the procedure is identical.
First we form the likelihood function:
We want the value of that maximizes this function, if it exists. If the domain of involves one or more endpoints, we know that the maximum might occur at one of the endpoints. However, the maximum might also occur at a point where the derivative of this function is zero, or where the derivative does not exist.
This requires differentiating the likelihood function. However, computing the derivative of a product with factors is not efficient. Since the logarithm (regardless of base) is an increasing function, any value of that maximizes the logarithm of will also maximize . Hence we take the natural logarithm to form the log-likelihood function:
Then we find all critical values of at which the derivative
After this we must use either the first derivative test or the second derivative test to determine whether or not there is a local maximum at any of the critical values or singular values we found. Comparing the value of the likelihood function at all endpoints, critical values and singular values allows us to determine the MLE of .
Note that the MLE is not necessarily an unbiased estimator. In fact, the MLE of is:
We now provide some examples. Note that, since , we will replace with whenever it occurs in the examples, which is handy for simplifying the computations.
Example 1: Suppose are the result of independent measurements on a continuous random variable whose probability distribution has the following pdf:
for some unknown parameter .
Find the MLE of .
Solution: The domain of is , which has no endpoints. The log-likelihood function is:
Then the derivative of the log-likelihood function is:
Moreover when , so we have one critical point.
Note that the second derivative
Consequently, the MLE of is .
Example 2: Suppose denote the result of independent measurements of a discrete random variable whose probability mass function is:
Find the MLE of .
Solution: The log-likelihood function is:
Then the derivative of the log-likelihood function is:
which equals when .
Note that:
- when we have so the derivative is positive and the log-likelihood function is increasing
- when we have so the derivative is negative and the log-likelihood function is decreasing
So we have a global maximum at .
There are no singular points and no other critical points, and the domain of has no endpoints.
Therefore the MLE of is .
Or visit omptest.org if jou are taking an OMPT exam.