Sufficient statistics

from Wikipedia, the free encyclopedia

In mathematical statistics one is sufficient statistic , also sufficient statistic called, is a statistic that all relevant information regarding the unknown parameter from the random sample contains. From the point of view of dimension theory, sufficiency with regard to a model is a possible property of measurable functions that map from the sample space into any measurement space . Such images are characterized as sufficient (also: exhaustive ) that transform a high-dimensional data vector into a simpler form without losing essential information about the underlying probability distribution . The counterpart of sufficiency is freedom of distribution ; it corresponds to an uninformative transformation.

Clearly formulated are precisely those statistics that contain all information about the parameters of the model to be estimated that are contained in the sample.

Sufficiency is one of the classic reduction principles of mathematical statistics, along with faithfulness to expectations and equivariance / invariance . Sufficiency gets its meaning from the theorem of Rao-Blackwell . It follows from this that "optimal" estimators with regard to the mean square error or corresponding generalizations can always be found in the set of sufficient estimators.

definition

Formally, the sample room , any measurement room and a measurable mapping between the two rooms. Furthermore, let be a random variable on the sample space, the distribution of which comes from a family of probability measures . then means sufficient for the family if the distribution does not depend on.

More generally, the sufficiency of a statistic can be defined by means of the sufficiency of σ-algebras : A statistic is called sufficient , or exhaustive if the σ-algebra it generates is a sufficient σ-algebra.

Example: binomial distribution

We get a simple example of sufficient statistics when examining identically independent Bernoulli-distributed random variables. The underlying model is therefore a Bernoulli process . In this case, the counting density of the random variable is given by, where they are either 0 or 1. Note that the counting measure is finite (especially -finite) and dominates the class because of the existence of the densities. Therefore one recognizes from the Neyman characterization that sufficient for is.

With the help of the definition, one shows the sufficiency of by calculating. If one now uses conditional probabilities one obtains:

.

This conditional density is now independent of and therefore sufficient.

So speaking heuristically, instead of knowing the entire data vector, it is sufficient to simply know the number of successes in this experiment in order to obtain all information about the unknown parameter .

Sentences about sufficiency in dominated distribution classes

Halmos-Savage theorem

Halmos-Savage's theorem provides a sufficiency criterion under the assumption that the distribution class is dominated. If countless infinite measures of the distribution class can then be combined into one measure , so that this dominates the distribution class and each probability measure of the distribution class has a measurable density , then it is a sufficient σ-algebra.

Neyman criterion

Assuming that there is a dominated distribution class , a statistic is sufficient if and only if measurable functions exist, so that the density can be broken down as follows: This characterization of the sufficiency goes back to Jerzy Neyman . In particular, bijective transformations of sufficient statistics are sufficient again. The Neyman criterion is derived from the Halmos-Savage theorem, but is easier to handle.

More sufficiency terms

Minimal sufficiency

Minimal sufficiency is a stronger requirement than sufficiency, which is also defined for statistics and σ-algebras. It poses the question of the maximum possible data compression, i.e. the smallest possible sufficient σ-algebra.

Strong sufficiency

Strong sufficiency is a modification of the conventional term sufficiency, which is defined using Markov kernels . For Borel spaces, strong sufficiency and sufficiency coincide.

Important sentences

Web links

literature

Individual evidence

  1. ^ Leonhard Held and Daniel Sabanés Bové: Applied Statistical Inference: Likelihood and Bayes. Springer Heidelberg New York Dordrecht London (2014). ISBN 978-3-642-37886-7 , p. 41.