Estimating the variance of an estimator

from Wikipedia, the free encyclopedia

As estimate of the variance of an estimation function in the statistics to estimate the variance of a is estimation function , an unknown parameter of the population, respectively. This estimation is a method of measuring the accuracy of estimation procedures. It allows the construction of confidence intervals ( interval estimation ).

If you have an estimate function for an unknown parameter of the population, you initially only have a point estimate for this. However, one is interested in also giving confidence intervals for the estimated parameter, i.e. H. one has to know the distribution and the variance of .

However, this is not always possible and therefore there are different procedures:

If the estimator was calculated using the maximum likelihood method , one knows about the asymptotic behavior:

  • such as

with the covariance matrix of the estimator (s) and the Fisher information matrix .

Known distribution of

If the distribution of can be determined at least approximately, for example with the aid of the central limit theorem , then the variance can easily be estimated.

An example is the sample mean of a normally distributed population or, if the central limit theorem is valid, for any distribution in the population:

.

The confidence interval can be derived from this

with from the standard normal distribution .

Direct procedure

The representation is used for direct methods

or multivariate

Variance estimates based on this can usually only be given for simple point estimates. Here, approximation formulas are only required for sample designs with second-order inclusion probabilities . Exact methods, i.e. formulas that are easy to calculate, can be given in the case of a linear estimator.

However, neither the true parameter nor the function are known. Therefore, the estimated values ​​and the normalized likelihood function are used as the probability density for :

or multivariate

The estimate is then made with the help of numerical integration .

Linear approximation

In the case of non-linear estimators (e.g. a ratio estimator ), approximate methods are used. Can the log-likelihood function be expanded around the maximum using the Taylor approximation

and taking advantage of the Fisher information matrix definition

follows

.

Alternatively, Woodruff linearization can be used to convert non-linear estimators to linear ones.

Resampling methods

Another possibility is resampling methods such as the bootstrapping process . Here, sub-samples are drawn at random from the existing sample and an estimated value is calculated with them. These estimates are an empirical approximation to the unknown distribution of .

Sample:
Subsample 1:
Sub-sample B: 

Hence it results

with . During the estimation, the sample design can be taken into account by weighting.