The maximum likelihood estimate is often used to estimate the variance of the population . The maximum likelihood estimate provides the uncorrected sample variance as an estimator of the unknown variance of the population , which, however, is only asymptotically true to expectations . An unbiased estimator, the corrected sample variance , is obtained by multiplying the uncorrected sample variance by the correction factor.
![{\ displaystyle 1 / (n-1)}](https://wikimedia.org/api/rest_v1/media/math/render/svg/0b0cf636940ea9df06c26ee645513e1cb712c34f)
Variance estimation of a normally distributed population
Maximum likelihood estimation
Let be independently and identically distributed random variables from a normally distributed population with the unknown expected value and the unknown variance of the population . Let the realizations of the random variables be , then the likelihood function (also called plausibility function) is a sample with size
![\ mu](https://wikimedia.org/api/rest_v1/media/math/render/svg/9fd47b2a39f7a7856952afec1f1db72c67af6161)
![x_ {1}, \ ldots, x_ {n}](https://wikimedia.org/api/rest_v1/media/math/render/svg/737e02a5fbf8bc31d443c91025339f9fd1de1065)
![{\ displaystyle L (x_ {1}, \ ldots, x_ {n} \ mid \ mu, \ sigma ^ {2}) = \ prod _ {i = 1} ^ {n} {\ frac {1} {\ sqrt {2 \ pi \ sigma ^ {2}}}} \ exp \ left (- {\ frac {(x_ {i} - \ mu) ^ {2}} {2 \ sigma ^ {2}}} \ right ) = \ left ({\ frac {1} {2 \ pi \ sigma ^ {2}}} \ right) ^ {n / 2} \ exp \ left (- {\ frac {1} {2 \ sigma ^ { 2}}} \ sum _ {i = 1} ^ {n} (x_ {i} - \ mu) ^ {2} \ right)}](https://wikimedia.org/api/rest_v1/media/math/render/svg/5fb962bae1311b06962b9cad38b73c27a0aeb77c)
and the log likelihood function
-
.
In order to find an estimator for , the log-likelihood function is derived from![{\ hat {\ sigma}} ^ {2}](https://wikimedia.org/api/rest_v1/media/math/render/svg/1ad9d89160c9e63c0aa4c158282cb75a894de56f)
![\ sigma ^ {2}](https://wikimedia.org/api/rest_v1/media/math/render/svg/53a5c55e536acf250c1d3e0f754be5692b843ef5)
![{\ displaystyle {\ frac {\ partial \ log (L (x_ {1}, \ ldots, x_ {n} \ mid \ mu, \ sigma ^ {2}))} {\ partial \ sigma ^ {2}} } = - {\ frac {n} {2 \ sigma ^ {2}}} + {\ frac {1} {2 \ sigma ^ {4}}} \ sum _ {i = 1} ^ {n} (x_ {i} - \ mu) ^ {2}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/05f2414d866a2f8501cd6423183a8c8425daa822)
and set equal to zero to find a maximum
![0 = - {\ frac {n} {2 {\ hat {\ sigma}} ^ {2}}} + {\ frac {1} {2 {\ hat {\ sigma}} ^ {4}}} \ sum _ {{i = 1}} ^ {n} (x_ {i} - \ mu) ^ {2} \ quad \ Longrightarrow \ quad {\ hat {\ sigma}} ^ {2} = {\ frac {1} {n}} \ sum _ {{i = 1}} ^ {n} (x_ {i} - \ mu) ^ {2}](https://wikimedia.org/api/rest_v1/media/math/render/svg/eedc572422b086f3a9049be37d7768cf820c8dde)
(for a derivation of the population variance in matrix notation, see classical linear model ). The second derivative results as
![{\ displaystyle {\ frac {\ partial ^ {2} \ log (L (x_ {1}, \ ldots, x_ {n} \ mid \ mu, \ sigma ^ {2}))} {\ partial \ sigma ^ {2} \ partial \ sigma ^ {2}}} = {\ frac {1} {\ sigma ^ {4}}} \ left ({\ frac {n} {2}} - {\ frac {\ sum _ {i = 1} ^ {n} (x_ {i} - \ mu) ^ {2}} {\ sigma ^ {2}}} \ right)}](https://wikimedia.org/api/rest_v1/media/math/render/svg/6cdf46e1153d8bc808726987f6a769d57ac9deda)
and at the point :
![\ sigma ^ {2} = {\ hat {\ sigma}} ^ {2}](https://wikimedia.org/api/rest_v1/media/math/render/svg/9a97cdc1ca6b324a9073b4a5f80820a3fdd738aa)
-
,
d. H. it is a maximum if .
![{\ hat {\ sigma}} ^ {2}> 0](https://wikimedia.org/api/rest_v1/media/math/render/svg/557b2d2acd30094e973df0453e5d5e4e83caa056)
Individual evidence
-
↑ Jürgen Hedderich, Lothar Sachs : Applied Statistics: Collection of Methods with R. , p. 332.