Mean Squared Error (MSE) vs. Mean Squared Logarithmic Error (MSLE): A Guide

Mean squared error (MSE) and mean squared logarithmic error (MSLE) are loss functions that significantly impact your data analyses. Here’s what you need to know.

Written by Mór Kapronczay
Published on Apr. 03, 2023
Mean Squared Error (MSE) vs. Mean Squared Logarithmic Error (MSLE): A Guide
Image: Shutterstock / Built In
Brand Studio Logo

Mean squared error (MSE) and mean squared logarithmic error (MSLE) are two loss functions that can significantly impact the results of your analyses. While MSE is one of the most common loss functions, it has some lesser known drawbacks. It works well with continuous targets, but it has some quirks with data in the 0-1 territory. Large outliers can also severely influence the loss, drawing away the focus.

On the other hand, MSLE isn’t used as often as MSE, but it solves some of the  shortcomings MSE has by utilizing the traits of a logarithm. It’s less prone to large outliers in the data set, and treats them as if they were broadly on the same scale. However, those outliers might be important to you.  

Mean Squared Error (MSE) vs. Mean Squared Logarithmic Error (MSLE): What’s the Difference?

  • Mean squared error (MSE): One of the most commonly used loss functions, MSE takes the mean of the squared differences between predicted and actual values to calculate your loss value for your prediction model. It works best when you’re doing a baseline analysis and you have a data set in a similar order of magnitude.
  • Mean squared logarithmic error (MSLE): MSLE takes a similar approach as MSE, but it utilizes a logarithm to off-set the large outliers in a data set and treats them as if they were on the same scale. This is most valuable if you aim for a balanced model with similar percentage errors.   

In an age of abundance of sophisticated deep learning models, it’s easy to overlook the basics. No doubt, adding another attention head to your neural network might contribute to your success considerably. However, there is much less talk about decisions that a practitioner has to make that can affect their results even more.

After all, a supervised machine learning problem in statistical terms is all about predicting a target variable based on a set of features, or independent variables. This prediction is judged based on a loss function, a metric quantifying how close your predicted values are to the actual target variable values. And this is not only important in terms of evaluation, during model training, a lot of decisions are taken automatically based on this loss (or objective) function. Selecting the right loss function is a critical choice to make, and one that practitioners often overlook despite its significance.

 

How to Select the Best Loss Function for Your Model

The loss function you select can have a significant impact on the model you’re building. 

Imagine that you’re working for a supermarket chain, and you want to predict the sales for individual products in its various points of sale. Most likely, sales distribution data will be skewed. Some stores will sell orders of magnitude more than others, and the same would apply to products. As a result, your target variable will have a power law distribution. To make matters worse, products have different shelf lives. So, to make a fair comparison, one might consider daily average sold quantities.

A well-known trick is to take the natural logarithm of the values, which forces the distribution to be closer to normal, in order to transform the target variable. Still, you’re dealing with a target variable with vast differences.

In these situations, the loss function you optimize for greatly influences your predictions. One might aim for the predictions to be in a certain percentile range of your targets. You could argue that “My predictions are in the ‘+-x% range’ of the sales numbers.” On the other hand, you might aim for actual sales number precision, as in, “My predictions are right with ‘x unit’ precision.” On top of that, you need to decide on whether you want to focus your predictions on the stores with lower revenue or the fewer high-achiever stores. These tradeoffs are mostly determined by which loss function you choose for your statistical model. 

So, how do you select the right loss function model?

More on Machine LearningHow Loss Functions Work in Neural Networks and Deep Learning

 

What Is Mean Squared Error (MSE)?

A popular option is to optimize using mean squared error (MSE). This involves minimizing the average of the squared errors, or taking the mean of the squared differences between predicted and actual values.

 

How Do You Calculate MSE? 

In Python, you’re most likely going to use the sklearn.metrics.mean_squared_error function. This function will take the actual true y values and your predicted ones, and it will return the value of the loss function. If you want to calculate it from scratch, you are going to need the formula:

mse equation
MSE equation. | Image: Mor Kapronczay

To calculate it, you’ll subtract the predicted values from the actual target values, square those differences 1-1 and then take the mean of the resulting squared error array.

import numpy as np

actual = np.array([1, 2, 3, 4, 5])
predicted = np.array([1.1, 1.9, 2.7, 4.5, 6])

def mse(actual: np.ndarray, predicted: np.ndarray) -> float:
    differences = np.subtract(actual, predicted)
    squared_differences = np.square(differences)
    return np.mean(squared_differences)

mse(actual, predicted)
# 0.27199999999999996

 

When Should You Use MSE?

The main thing you have to consider when deciding your loss function is how your target variable looks, and what error distribution you would tolerate more. 

MSE transformation line graph
MSE target variable transformation graph. | Image: Mor Kapronczay 

Small values, such as values under one, would contribute nearly nothing to the metric. Therefore, the optimization more or less disregards them, resulting in relatively large errors. It’s important to note that squaring the errors lowers the significance of nominal errors under one and amplifies those over one. Consequently, large errors drive the loss, so that the algorithm focuses more on large values than small ones.

In summary, you should consider mean squared error if:

  • You are doing a baseline analysis before making any decisions.
  • You have some star performers, and you are the most keen on making a good model for those.
  • Your data is predominantly in the same order of magnitude.
A tutorial on mean squared error. | Video: Simplilearn

 

What Is Mean Squared Logarithmic Error (MSLE)?

Mean squared logarithmic error (MSLE) is a less commonly used loss function. It’s considered to be an improvement over using percentage based errors for training because its numerical properties are better, but it essentially serves the same purpose: Trying to create a balance between data points with orders of magnitude difference during model training.

 

How Do You Calculate MSLE?

In Python, you most probably are going to use sklearn.metrics.mean_squared_logarithmic_error, which works exactly like the MSE counterpart. It’s easy to calculate it from scratch using code: 

MSLE equation
MSLE equation. | Image: Mor Kapronczay

You have to add one to the actual and predicted target values, and take their differences by subtracting the latter from the former. Subsequently, we square those logarithmic differences 1-1, then take the mean.

import numpy as np

actual = np.array([1, 2, 3, 4, 5])
predicted = np.array([1.1, 1.9, 2.7, 4.5, 6])

def msle(actual: np.ndarray, predicted: np.ndarray) -> float:
    log_differences = np.subtract(np.log(1 + actual), np.log(1 + predicted))
    squared_log_differences = np.square(log_differences)
    return np.mean(squared_log_differences)

msle(actual, predicted)

 

When Should You Consider MSLE?

The logarithm essentially makes the error profile more flat, reducing the impact of the larger values. The plus one evades the logarithm, producing negative values for errors between zero and one. It’s important to note that this metric assumes non-negative target variables. As a result, the previously seen overwhelming effect of large values is reduced, producing a more equal emphasis on data points.

MSLE target variable transformation line graph
MSLE target variable transformation line graph. | Image: Mor Kapronczay

In summary, you should use MSLE if:

  • You want to level the playing field for target values that have different orders of magnitudes.
  • You aim for a balanced model having roughly similar percentage errors.
  • You can tolerate large differences in terms of units for large target values.

More on Data ScienceL1 and L2 Regularization, Methods Explained

 

Mean Squared Error (MSE) vs. Mean Squared Logarithmic Error (MSLE)

Back to the store example. Let’s run a quick simulation predicting the sales of individual products for the next period using sales data from the current (previous) period as a feature. We’ll look at two models, one that’s trained with mean squared error (MSE) and the other mean squared logarithmic error (MSLE). The bars show the percentage difference between the results of the two models in terms of their respective error bases.

mse vs msle bar graph comparison
A bar graph comparison of the error metrics for MSE vs MSLE. | Image: Mor Kapronczay

 Looking at the graph, the negative percentage differential shows that the MSE model is more effective in that particular segment, producing smaller errors, while the positive differential indicates that the MSLE model is performing better for that group. 

For instance, the first bar group shows large sales places/products. The MSE objective model has a 30 percent better percentage error and is 56 percent better when errors are measured in sales units. On the other hand, when looking at small sales, the MSLE model is better in terms of units and percentages. The same phenomenon can be seen on the third bar group. Percentage-wise, MSLE performs better because of the better percentage errors it produces for small sales, which make up most of the data. Meanwhile, MSLE optimization results in large errors in sales units for large sales, effectively making MSE a slightly better performer in terms of units over the whole group. 

So, what should you learn from all of this? In my view, these are the most important takeaways from this chart:

  1.  MSE trained models perform better on large sales occasions. These are generally fewer but might be more important. In contrast, MSLE performs better for the average, small sales stores.
  2. In all data subsets, MSLE models provide an improvement if errors are measured in percentages. The same applies for MSE if the errors are in sales units.
  3. Always consider the loss function you want to optimize for in your use case. Don’t just go with the default one.
Hiring Now
Outdoorsy
Insurance • Sharing Economy • Software • Travel
SHARE