Multi-dimensional normal distribution

from Wikipedia, the free encyclopedia
Density of the bivariate normal distribution in three-dimensional space

The multidimensional or multivariate normal distribution is a multivariate distribution In the multivariate statistics . It represents a generalization of the (one-dimensional) normal distribution over several dimensions . A two-dimensional normal distribution is also called a bivariate normal distribution .

A multi-dimensional normal distribution is determined by two distribution parameters - the expected value vector and the covariance matrix , which correspond to the parameters ( expected value ) and ( variance ) of the one-dimensional normal distributions.

Multi-dimensional, normally distributed random variables appear as limit values ​​of certain sums of independent multi-dimensional random variables. This is the generalization of the central limit theorem for the multi-dimensional central limit theorem .

Because they occur where multidimensional random quantities can be viewed as a superposition of many independent individual effects, they are of great importance in practice.

Due to the so-called reproductive property of the multi-dimensional normal distribution, the distribution of sums (and linear combinations ) of multi-dimensional normally distributed random variables can be specified in concrete terms.

The multidimensional normal distribution: regular case

10000 samples of a two-dimensional normal distribution with , and ρ = 0.7

A -dimensional real random variable is multidimensionally normally distributed with an expected value vector and (positively definite, i.e. regular ) covariance matrix if it is a density function of the form

owns. One writes

.

The subscript is the dimension of the -dimensional normal distribution and indicates the number of variables; i.e., is , is and is . There is no closed form for the associated distribution function . The corresponding integrals must be calculated numerically.

The value in the exponential part of the density function corresponds to the Mahalanobis distance , which represents the distance from the test point to the mean value . In comparison with the density function of the one-dimensional normal distribution, the covariance matrix plays the role of the scalar unknown variance in the multi-dimensional normal distribution .

properties

The multidimensional normal distribution has the following properties:

  • The affine transformation with a matrix (with ), and is normally distributed -dimensional: . According to the definition given here, this only applies if is non-singular, i.e. has a non-vanishing determinant.
  • The affine transformation
standardizes the random vector : it is (with identity matrix ).
  • Conditional distribution with partial knowledge of the random vector: If a multidimensional, normally distributed random vector is conditional on a sub-vector, the result itself is again multidimensional, normally distributed, for
applies
,
in particular, the expected value depends linearly on the value of and the covariance matrix is ​​independent of the value of .

The multidimensional normal distribution: general case

If the covariance matrix is singular, you can not invert, then there is no density in the form given above. At the same time, one can also define the multidimensional normal distribution, but now using the characteristic function .

A -dimensional real random variable is called normally distributed with an expected value vector and (positive semidefinite, i.e. not necessarily regular) covariance matrix if it has a characteristic function of the following form:

.

If is regular, there is a probability density in the above form, if is singular, then in -dimensional space there is no density with respect to the Lebesgue measure . Suppose , then there is a -dimensional linear form , where is a -Matrix, which satisfies a -dimensional normal distribution with an existing density im .

The marginal distribution of the multidimensional normal distribution

Bivariate normal distribution with marginal distributions

Let be multidimensional normally distributed. For any partition with and , , holds that the marginal distributions and (multi-dimensional) normal distributions are.

However, the reverse does not apply, as the following example shows:

Be and be defined by

whereby . Then is also and

.

Accordingly, the covariance (and thus the correlation coefficient ) of and is zero if and only if . From the non-correlation of two random variables , and it would be normally distributed multidimensional case immediately following independence (special feature of the multivariate normal distribution), but since and , by definition, are not independent ( always the same ), may in particular not be normally distributed multidimensional.

The p -dimensional standard normal distribution

The measure of probability based on that given by the density function

is defined is called the standard normal distribution of the dimension . The dimensional standard normal distribution is apart from translations (d. H. Expectation value ) is the only multidimensional distribution whose components are stochastically independent and their density also rotationally symmetrical is.

Moments and accumulators

As in the one-dimensional case, all moments of the multi-dimensional normal distribution are defined by the first two moments. All cumulants except the first two are zero. The first two cumulants are the expected value vector and the covariance matrix . With regard to the multidimensional moment problem , the normal distribution has the property that it is uniquely defined by its moments. That is, if all the moments of a multidimensional probability distribution exist and correspond to the moments of a multidimensional normal distribution, the distribution is the unambiguous multidimensional normal distribution with these moments.

Density of the two-dimensional normal distribution

The density function of the two-dimensional normal distribution with mean values and and correlation coefficient is

10,000 samples each of two-dimensional normal distributions with ρ = −0.8, 0, 0.8 (all variances are 1).

In the two-dimensional case with mean values and any variances, the density function is

The general case with arbitrary means and variances is obtained through translation (replace through and through )

Example of a multidimensional normal distribution

An apple tree plantation is considered with a large number of apple trees of the same age, i.e. comparable ones. One is interested in the characteristics of the size of the apple trees, the number of leaves and the yields. So the random variables are defined:

: Height of a tree [m]; : Yield [100 kg]; : Number of sheets [1000 pieces].

The variables are each normally distributed like

.

Most trees are therefore large, very small or very large trees are rather rare. A large tree tends to have a higher yield than a small tree, but of course there will be a large tree with little yield every now and then. Yield and size are correlated, the covariance is and the correlation coefficient .

It is the same with the correlation coefficient , and with the correlation coefficient .

If the three random variables are combined in the random vector , the distribution is multidimensional. However, this does not apply in general (cf. the marginal distribution of the multidimensional normal distribution ). In the present case then applies to the joint distribution of

and

The corresponding sample correlation matrix is

Estimation of the parameters of the multidimensional normal distribution

In reality, the distribution parameters of a -dimensional normal distribution will usually not be known. So these parameters need to be estimated .

You take a sample of the size . Every realization of the random vector could be understood as a point in a -dimensional hyperspace. This gives the one matrix (test plan or data matrix ):

, in which

which contains the coordinates of a point in each line (see multiple linear model in matrix notation ).

The expected value vector is estimated by the mean value vector of the arithmetic mean values ​​of the columns of

with the components

.

This estimator is the best unbiased estimator for the expected value vector with regard to the mean square deviation . However, it is deemed not to be permissible in the sense of decision theory . Then there are better estimators, e.g. B. the James Stein estimator .

The data matrix centered with respect to the arithmetic mean values ​​proves useful for estimating the covariance matrix . It is calculated as

,

with the elements , where the one vector represents a column vector of length with all ones. The arithmetic mean of the associated column is subtracted from all entries.

The covariance matrix has the estimated components

.

It turns out to be

.

The population correlation matrix is estimated by the pairwise correlation coefficients

,

on their main diagonal there are ones.

Sample to sample

10 apple trees were randomly selected and 3 properties were measured in each case:: Height of a tree [m]; : Yield [100 kg]; : Number of sheets [1000 pieces]. These observations are summarized in the data matrix :

.

The mean values ​​are calculated, as shown by way of example , as

.

They give the mean value vector

.

The centered observations are obtained for the centered data matrix by subtracting the corresponding mean value from the columns:

,

so

.

The covariances are calculated for the covariance matrix, as in the example,

and accordingly the variances

,

so that the sample covariance matrix

results.

Correspondingly, for example , one obtains for the sample correlation matrix

or in total

.

Generation of multi-dimensional, normally distributed random numbers

An often used method for generating a random vector of a -dimensional normal distribution with an expected value vector and ( symmetric and positively definite ) covariance matrix can be given as follows:

  1. Find a matrix such that . The Cholesky decomposition of or the square root of can be used for this.
  2. Let be a vector whose components are stochastically independent, standard normally distributed random numbers. These can be generated using the Box-Muller method , for example.
  3. The desired -dimensional normal distribution results from the affine transformation .

Scatter regions of the multidimensional normal distribution

For one-dimensional normally distributed random variables, approximately 68.27% of the realizations lie in the interval ; for multi-dimensional normally distributed random variables, the regions of constant probability are given by ellipses (the standard deviation ellipses ) which are centered around the mean. The main axes of the ellipse are given by the eigenvectors of the covariance matrix , the length of the semi-axes is the square root of the eigenvalue belonging to the respective main axis . Finding a realization of the random variables in the region which is limited by the (multidimensional) standard deviation ellipse is less likely for a multidimensional, normally distributed random variable.

Representation of the standard deviation ellipse of a two-dimensional normal distribution, as well as the two marginal distributions .

After a main axis transformation, the axes can be normalized with their respective ones. Then the probability can be calculated as a function of the radius with which a measured value lies within this radius. With

is the share

of the measured values ​​at most at a distance from the mean value of a p-dimensional normal distribution. It is the regularized incomplete gamma function of the upper limit.

in %
68.27 95.45 99.73
39.35 86.47 98.89
19.87 73.85 97.07

Correspondingly, the inverse function can be used to specify the scattering radius r, in which a specified proportion of measured values ​​lies:

in
0.675 1.645 2.576
1.177 2.146 3.035
1.538 2,500 3.368

literature

  • Mardia, KV, Kent, JT, Bibby, JM: Multivariate Analysis , New York 1979
  • Fahrmeir, Ludwig, Hamerle, Alfred, Tutz, Gerhard (Eds.): Multivariate statistical methods , New York 1996
  • Hartung, Joachim, Elpelt, Bärbel: Multivariate Statistics , Munich, Vienna 1999
  • Flury, Bernhard, A first course in multivariate statistics , New York, 1997.

Remarks

  1. Multidimensional and multivariate normal distributions are used synonymously in this article. In Hartung / Elpelt: Multivariate statistics they have different meanings (in Chapter 1, Section 5): Here the multivariate normal distribution is a matrix distribution.
  2. ^ Rencher, Alvin C., and G. Bruce Schaalje: Linear models in statistics. , John Wiley & Sons, 2008, p. 89.
  3. ^ Rencher, Alvin C., and G. Bruce Schaalje: Linear models in statistics. , John Wiley & Sons, 2008, p. 90.
  4. Kleiber, Stoyanov: Multivariate distributions and the moment problem, Journal of Multivariate Analysis , Volume 113, January 2013, pages 7–18, doi: 10.1016 / j.jmva.2011.06.001 .
  5. Bin Wang, Wenzhong Shi, Zelang Miao: Confidence Analysis of Standard Deviational Ellipse and Its Extension into Higher Dimensional Euclidean Space . In: PLOS ONE . tape 10 , no. 3 , March 13, 2015, ISSN  1932-6203 , p. 11 , doi : 10.1371 / journal.pone.0118537 .
  6. would be described in the normalization .