Goodness Measure

What makes a 'good' forecast? The answer to this central question depends heavily on the underlying application case. In statistics, a wide variety of goodness measures are used to evaluate the quality of a forecast. This already suggests that there is no single, optimal goodness measure for evaluating forecasts. Rather, when assessing goodness measures, the knowledge about the data basis and the requirements for the forecast must be considered. Therefore, it is crucial to understand the principles on which these goodness measures are based.

Below, we will look at some of the most commonly used goodness measures:

What these goodness measures have in common is that their evaluation is based on the so-called forecast error.

Forecast Error

The forecast error (engl. forecast error) is the difference between the actual value that occurred and the forecasted value.

Specifically, the forecast error for a forecast looking i time units ahead is given by:

e t = act t - fc t

Where

fc t

is the forecasted value after i time units, and

act t

is the actual value that occurred at the corresponding time.

Properties

Forecasts and forecast errors of a monthly time series during the testing period

Absolute Error (AE)

The absolute value of the deviation from the actual value to the predicted value is called the absolute forecast error (engl. absolute forecast error, AE). It is calculated as:

AE t = | e t | = | act t - fc t |

Properties

Percentage Error (PE)

The percentage forecast error (engl. percentage error, PE) measures the forecast error relative to the actual value that occurred. It is given by:

PE t = e t act t = act t - fc t act t

Properties

Absolute Percentage Error (APE)

The absolute percentage forecast error (engl. absolute percentage error, APE) measures the magnitude of the forecast error relative to the actual value. It is given by:

APE t = | e t act t | = | act t - fc t act t |

Apart from being restricted to non-negative values, the APE shares the properties of the percentage forecast error.

Properties

Mean Error (ME)

The mean error (engl. mean error, ME) represents the average error. The mean is usually calculated over all forecast steps from 1 to h.

ME = 1 h t = 1 h e t = 1 h t = 1 h ( act t fc t )

With the ME, we have arrived at the aggregated quality measures.

Properties

Mean forecast error (ME) of a time series in the test period

Mean Absolute Error (MAE)

The mean absolute error (MAE) gives the average of all absolute forecast errors. The average is usually calculated over all forecast steps from 1 to n.

MAE = 1 h t = 1 h AE t = 1 h t = 1 h e t = 1 h t = 1 h act t fc t

Apart from the fact that forecast errors cannot cancel each other out due to the restriction to non-negative values, the MAE shares the properties of the mean error.

Properties

Mean absolute forecast error (MAE) of a time series in the test period

Mean Absolute Percentage Error (MAPE)

The mean absolute percentage error (MAPE) describes the average of the absolute percentage forecast errors relative to the magnitude of the actual values.

MAPE = 1 h t = 1 h APE t = 1 h t = 1 h e t act t = 1 h t = 1 h act t fc t act t

Because division by zero would occur, the MAPE cannot be used for time series with many zero values. It also doesn't provide good results for time series with many values close to zero. However, unlike the MAE or MSE, the MAPE is unitless and can therefore be more useful for comparing the quality of forecasts of different magnitudes.

Here is the translation of your latest input into English, keeping the format as close as possible to the original:

Properties

Mean Absolute Percentage Forecast Error (MAPE) of a time series in the test period

Mean Squared Error (MSE)

The mean squared error (MSE) corresponds to the average of the squared forecast errors.

MSE = 1 h t = 1 h e t 2 = 1 h t = 1 h ( act t fc t ) 2

Like the MAE and MAPE, the MSE only takes into account the absolute deviation of the forecast from the actual value, not its direction. In comparison to the MAE, large errors carry more weight due to the squaring. As a result, the MSE is more sensitive to outliers. The mean squared error is often used as an optimization criterion in model building, such as in classical linear regression.

Properties

Mean Squared Forecast Error (MSE) of a time series in the test period

Mean Absolute Scaled Error (MASE)

The mean absolute scaled error (MASE) corresponds to the MAE of the considered forecast divided by the MAE of a one-step naive forecast (in-sample) of the actual values from 1 to n.

MASE = 1 h t = 1 h act n+t fc n+t 1 n-1 t = 2 n act t act t-1

Therefore, a MASE greater than 1 implies that the considered forecast is worse than a one-step naive forecast; a MASE less than 1 implies that it is better. While a good one-step forecast should clearly have a MASE below 1, for a multi-step forecast, a MASE greater than 1 does not necessarily mean that the forecast is not good.

Like the MAPE, the MASE has no unit and is therefore suitable for comparisons. Compared to the MAPE, the MASE can handle (individual) zero values in time series better. However, the MASE is not well suited for nearly constant time series, as in this case the forecast errors of a naive forecast are often zero, making its MAE very small.

Properties

Symmetric MAPE (sMAPE)

The symmetric mean absolute percentage error (sMAPE) averages the absolute errors divided by the mean of the absolute values of the actual and forecasted values.

sMAPE = 1 h ( | e t | ( | act t | + | fc t | ) / 2 ) = 1 h ( | act t - fc t | ( | act t | + | fc t | ) / 2 )

Compared to the MAPE, where the weighting of the forecast errors is based only on the actual value, the sMAPE also takes into account the magnitude of the forecasted value. Like the MAPE, the sMAPE does not yield good results when many of the actual or forecasted values are near or equal to zero.

The sMAPE takes values between 0% and 200%.

Properties

MAPE vs. sMAPE

The transformed errors, which when averaged result in the MAPE or sMAPE, behave quite similarly; the difference between both metrics is most clearly shown in the following example. For simplicity, only one forecast time point is considered.

Case 1: The forecast is fixed at fc = 100. The actual values vary around 100 +/- 10. This leads to an absolute forecast error of 10 in both cases.

act = 90, fc = 100 act = 110, fc = 100
MAPE 10/90 10/110
sMAPE 10/95 10/105

Case 2: Now, we fix the actual values at act = 100 and vary the forecast around 100 +/- 10. As in Case 1, the forecast error is 10 in both cases.

act = 100, fc = 90 act = 100, fc = 110
MAPE 10/100 10/100
sMAPE 10/95 10/105

One can observe:

Periods in Stock (PIS)

Periods in Stock is a metric that sums up how long forecast errors stay as stock in a hypothetical warehouse before they are balanced out by corresponding forecast errors in the opposite direction.

PIS = - t = 1 h j = 1 t e j = - t = 1 h j = 1 t ( act j - fc j )

The direction of the forecast errors is important for this metric. PIS, unlike other metrics such as MAE, takes into account the duration of mismatches between the forecast and the actual value and is therefore well-suited for evaluating forecasts of sporadic time series, i.e., time series with many zero values.

For example, a forecast several days too early (Forecast 1) results in a higher PIS, i.e., a worse forecast quality, compared to a forecast that is only one day shifted (Forecast 2), while other metrics like MAE would evaluate both cases equally.

Time Point Actual Value Forecast 1 "Stock" 1 Forecast 2 "Stock" 2
1 0 1 1 0 0
2 0 0 1 0 0
3 0 0 1 1 1
4 1 0 0 0 0

PIS1 = 3,   PIS2 = 1,   MAE1 = MAE2 = 0.5

Characteristics

You are about to leave our website via an external link. Please note that the content of the linked page is beyond our control.

Cookies und andere (Dritt-)Dienste

Diese Website speichert Cookies auf Ihrem Computer nur, wenn Sie dem ausdrücklich zustimmen. Bei Zustimmung werden insbesondere auch Dritt-Dienste eingebunden, die zusätzliche Funktionalitäten, wie beispielsweise die Buchung von Terminen, bereitstellen. Diese Cookies und Dienste werden verwendet, um Informationen darüber zu sammeln, wie Sie mit unserer Website interagieren, und um Ihre Browser-Erfahrung zu verbessern und anzupassen. Zudem nutzen wir diese Informationen für Analysen und Messungen zu unseren Besuchern auf dieser Website und anderen Medien. Weitere Informationen zu den von uns verwendeten Cookies und Dritt-Diensten finden Sie in unseren Datenschutzbestimmungen.