HOME

TheInfoList



OR:

In statistics, the mean signed difference (MSD), also known as mean signed deviation and mean signed error, is a sample statistic that summarises how well a set of estimates \hat_i match the quantities \theta_i that they are supposed to estimate. It is one of a number of statistics that can be used to assess an estimation procedure, and it would often be used in conjunction with a sample version of the mean square error. For example, suppose a linear regression model has been estimated over a sample of data, and is then used to extrapolate predictions of the dependent variable out of sample after the out-of-sample data points have become available. Then \theta_i would be the ''i''-th out-of-sample value of the dependent variable, and \hat_i would be its predicted value. The mean signed deviation is the average value of \hat_i-\theta_i.


Definition

The mean signed difference is derived from a set of ''n'' pairs, ( \hat_i,\theta_i), where \hat_i is an estimate of the parameter \theta in a case where it is known that \theta=\theta_i. In many applications, all the quantities \theta_i will share a common value. When applied to
forecasting Forecasting is the process of making predictions based on past and present data. Later these can be compared (resolved) against what happens. For example, a company might estimate their revenue in the next year, then compare it against the actual ...
in a
time series analysis In mathematics Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in m ...
context, a forecasting procedure might be evaluated using the mean signed difference, with \hat_i being the predicted value of a series at a given lead time and \theta_i being the value of the series eventually observed for that time-point. The mean signed difference is defined to be :\operatorname(\hat) = \frac\sum^_ \hat - \theta_ .


See also

*
Bias of an estimator In statistics, the bias of an estimator (or bias function) is the difference between this estimator's expected value and the true value of the parameter being estimated. An estimator or decision rule with zero bias is called ''unbiased''. In s ...
*
Deviation (statistics) In mathematics and statistics, deviation is a measure of difference between the observed value of a variable and some other value, often that variable's mean. The sign of the deviation reports the direction of that difference (the deviation is posi ...
* Mean absolute difference *
Mean absolute error In statistics, mean absolute error (MAE) is a measure of errors between paired observations expressing the same phenomenon. Examples of ''Y'' versus ''X'' include comparisons of predicted versus observed, subsequent time versus initial time, and ...
Summary statistics Means Distance {{statistics-stub