M-estimator

from Wikipedia, the free encyclopedia

M-estimators , also maximum likelihood-like estimators, represent a class of estimation functions that can be viewed as a generalization of the maximum likelihood method . Compared to other estimators such as B. the maximum likelihood estimators more robust against outliers .

This article deals with M-estimators for determining the location parameter .

Derivation by generalization of the maximum likelihood method

The principle of maximum likelihood estimators is based on the function

with the corresponding density or probability function depending on .

The idea with M-estimators is to replace the function with a function that is less sensitive to outliers. The job is the expression

depending on to minimize or the equation

With

to solve.

Each solution to this equation is called an M-estimator.

Implicit definition

Let an arbitrary distribution function and an odd and monotonically increasing function not equal to 0. Then is defined as the solution of the equation

It should be noted that, depending on the choice of and, there can be either none, one or more solutions. In the case of a specific sample , the solution is

Called M-estimator.

Suitable functions ρ

The following are the according to

standardized to achieve scale invariance . represents a dispersion estimator for which the MAD (Median Absolute Deviation) is mostly used.

method
Least Squares Method
Huber k estimator
Fine art estimator
Andrews wave
Tukey's biweight

The weight functions in the following figure show the differences between the estimators: with Huber-k, even extreme observations have a low weight, with the Hampel, Andrews wave and Tukey's biweight estimators, extreme observations are assigned the weight zero.

Weight functions w (z) for different M-estimators. The parameter values ​​correspond to the standard values ​​of SPSS.

robustness

With a suitable choice of (even, bounded and monotonically increasing), M-estimators have a breakpoint of

Numerical solution method

For many functions no explicit solution can be given, so it has to be calculated numerically. As usual for calculating zero problems, the Newton-Raphson method is also available here , and the following iteration rule results, where again  :

The median is usually used as a suitable starting value . This iteration method converges very quickly, usually two to three iteration steps are sufficient.

W estimator

W-estimators are very similar to M-estimators and usually deliver the same results. The only difference is in the solution to the minimization problem. W-estimators are mostly used in robust regression .

It becomes the weighting function

With

introduced, with the help of which the minimization problem can be rewritten in

Inserting the definition of , multiplying and rearranging finally results in the fixed point equation

the iteration rule

See also

literature

  • Robert G. Staudte: Robust estimation and testing . Wiley, New York 1990. ISBN 0-471-85547-2
  • Rand R. Wilcox: Introduction to robust estimation and hypothesis testing . Academic Press, San Diego Cal 1997. ISBN 0-12-751545-3