Maximum likelihood estimation (MLE) is a method we use to estimate the parameters of a model so those chosen parameters maximize the likelihood that the assumed model produces the data we can observe in the real world.

What Is Maximum Likelihood Estimation Used For?

We can use MLE to estimate the parameters of regression models such as linear, logistic and Poisson regressions. We use these models in economics, finance and public health to analyze relationships between variables. We can also use MLE to estimate the parameters of more complex models, such as neural networks and decision trees.

A model is a mathematical representation of a process that produces our observed data. For example, we may use a linear model, which we can write as y=mx + c, to predict regional house prices. In this example, let’s say y equals the house prices we observe and x is equal to the number of rooms as an independent variable. This linear model describes the house prices in terms of the number of bedrooms. Parameters m and cact like the model’s blueprints. As you can see below, the characteristics of the model shift with the model’s changing parameters.

maximum likelihood estimation (MLE) image of a linear model demonstrating housing prices.
Linear models of house prices. | Image by Erdem Isbilen

With the help of MLE, we choose parameter values so that the model describes a process that’s most likely to produce the observed data.

More on Data Science and Statistics From Built In ExpertsWhat Is Pattern Recognition?

 

Best Practices of MLE in 3 Steps

  1. Assume a model for the observed data. The results will be heavily dependent on the model assumption so this is the most important step. 
  2. Calculate the joint likelihood function containing the likelihood functions of each data point in terms of the model parameters.
  3. Find the parameter values that maximize the joint likelihood function. To do this, we need to find the derivative of the likelihood function and set it to zero.

 

Properties of Maximum Likelihood Estimation

Consistency 

MLE is a consistent estimator, which means that as the sample size increases, the estimates obtained by the MLE approach the true values of the parameters if some conditions are met. In other words, as the sample size increases, the probability of getting the correct estimate of the parameters increases.

Related Reading From Built In ExpertsWhat Is the Law of Large Numbers?

 

Asymptotic Normality 

MLE follows the asymptotic normality property, which means that as the sample size increases, it converges in distribution to a normal distribution. MLE distribution here is the distribution of the parameter estimated with the different sets of observed data. This property is useful in hypothesis testing and constructing confidence intervals.

 

Invariance

The invariance property of MLE states that if we apply a function or transformation to the parameter we’re estimating, the MLE of the transformed parameter is equal to the transformation of the MLE of the original parameter. In other words, if we have an MLE for a parameter θ and we transform θ using a function g(θ), then the MLE of g(θ) is simply g(θ̂), where θ̂ is the MLE of θ.

Maximum Likelihood, Clearly Explained. | Video: StatQuest With Josh Starner

 

Maximum Likelihood Estimation Examples

Let’s continue with the house prices example above. 

maximum likelihood estimation (MLE) image of a linear model demonstrating housing prices.
Linear models of house prices. | Image by Erdem Isbilen

We have given sets of observed data that are shown as black dots in the above graph. They are(1, 2), (2, 6), (3,7), (4, 10) and (5,12). The numbers represent the bedroom number and the house price respectively.

We assume a linear model so we have two alternative solutions represented with two sets of parameters:

  1. (m:2, c:1)
  2. (m:4, c:-3) 

The red line represents the first alternative, which is modeled as y=2x+1 while the green line represents the second alternative as y=4x-3

We’ll use the MLE method to find which set of parameters provides a model that better fits the observed data.

Below you’ll find the observed data points in the xand ycolumns. 

y_est is the value our first model estimates. It’s the value we get when we plug the x values into the y=2x+1 linear equation. 

maximum likelihood estimation (MLE) table laying out the values as described by the author
Likelihood values of y = 2x + 1 function. | Image by Erdem Isbilen

To calculate the likelihood of a specific data point, we need to know the probability distribution of our data at each location. This means that we need to know (or assume) the distribution of house prices for each number of bedrooms. You can use any distribution function but for the sake of simplicity, I’ll use the normal distribution with a standard deviation of one.

Let’s see below how to calculate the likelihood value . Here’s the formula of the normal distribution that we consider for our data.

maxiumum likelihood estimation (MLE) image of a formula as described by author

In this formula: 
maximum likelihood estimation (MLE) is equal to
1

 Maximum likelihood estimation  is equal to (y -yest)2 as shown in the table above

When we plug the values into the probability distribution formula above, we can find the likelihood values of the data point.

So the likelihood value is 0.39 as seen in the distribution curve, which is the maximum possible value. 

Let’s look at another data point. For the observed data point of (4, 10), the estimated value by our model is 9 whereas the observed value is 10. This time, the likelihood value of the data point is 0.25 as you can see in the distribution curve below.

maximum likelihood estimation graph as described by author
Probability distributions of data points. | Image by Erdem Isbilen

You can see the same table for the other alternative model below. Likelihood values are small compared to the first model. This time, we have different y_est values in the table because they’re calculated by the y = 4x - 3 linear equation.

maximum likelihood estimation (MLE) table as described by author
Likelihood values of y = 4x - 3 function. | Image by Erdem Isbilen

As a result, likelihood values deteriorate as y_estvalues move away from the center of the distribution curve. For the data point (4,10), the likelihood value is almost zero because our model estimates the house price as 13 while the observed value is 10. The estimated value of 13 is far away from the distribution limits we assumed for the data point as you can see in the below graph.

maximum likelihood estimation (MLE) graph as described by author
Probability distributions of data points. | Image by Erdem Isbilen

That means the first model better fits the observed data.

Find out who's hiring.
See all Data + Analytics jobs at top tech companies & startups
View 2905 Jobs

 

Maximum Likelihood Estimation Advantages and Disadvantages 

Advantages of MLE

  1. MLE is known to be an efficient estimator, which means it produces estimates that have lower variances compared to other methods under certain assumptions.
  2. Asymptotically, MLE estimates become consistent as the sample size grows, which means that they converge to the true parameter values with probability 1.
  3. Under certain conditions MLE can produce unbiased estimates of the population parameters.
  4. We can apply MLE to a wide range of statistical models. It’s widely used in fields such as economics, finance, biology and engineering.

More Data Science Tutorials From Built In ExpertsPolynomial Regression: An Introduction

 

Disadvantages of MLE

  1. MLE is sensitive to outliers, which can significantly affect our estimates of the parameters.
  2. MLE requires knowledge of the probability distribution that generates the data. If the distribution is misspecified, the estimates can be biased or inconsistent.
  3. In some cases, the likelihood function may be difficult to maximize, which can make the estimation process computationally expensive.
  4. MLE requires large sample sizes for accurate estimation but in some cases, you may not be able to collect large samples.
Expert Contributors

Built In’s expert contributor network publishes thoughtful, solutions-oriented stories written by innovative tech professionals. It is the tech industry’s definitive destination for sharing compelling, first-person accounts of problem-solving on the road to innovation.

Learn More

Great Companies Need Great People. That's Where We Come In.

Recruit With Us