The goal of S-estimators is to have a simple
high-breakdown regression estimator, which share the flexibility and nice asymptotic properties of
M-estimators. The name "S-estimators" was chosen as they are based on estimators of scale.
We will consider estimators of scale defined by a function
, which satisfy
* R1 –
is symmetric,
continuously differentiable
In mathematics, a differentiable function of one real variable is a function whose derivative exists at each point in its domain. In other words, the graph of a differentiable function has a non- vertical tangent line at each interior point in ...
and
.
* R2 – there exists
such that
is strictly increasing on
For any sample
of real numbers, we define the scale estimate
as the solution of
,
where
is the
expectation value of
for a
standard normal distribution
In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is
:
f(x) = \frac e^
The parameter \mu i ...
. (If there are more solutions to the above equation, then we take the one with the smallest solution for s; if there is no solution, then we put
.)
Definition:
Let
be a sample of regression data with p-dimensional
. For each vector
, we obtain residuals
by solving the equation of scale above, where
satisfy R1 and R2. The S-estimator
is defined by
and the final scale estimator
is then
.
[P. Rousseeuw and V. Yohai, Robust Regression by Means of S-estimators, from the book: Robust and nonlinear time series analysis, pages 256–272, 1984]
References
Estimator
Robust regression