Cumulative

from Wikipedia, the free encyclopedia

In probability theory and statistics, cumulants are parameters of the distribution of a random variable which, with regard to the summation of stochastically independent random variables, satisfy simple arithmetic laws. The sequence of cumulants begins with the expected value and the variance .

definition

Is the moment generating function of the random variable , i. h., it is

,

that's the name of the function

Cumulant generating function . The -th cumulant of the distribution of is then defined by

.

Alternatively, cumulants can also be defined by the characteristic function of a random variable .

The -th cumulant is then defined by

properties

Displacement invariance

The cumulants are also referred to as semi-invariants of the density function because, with the exception of , they do not change when the expected value is shifted. Let it be a random variable, then for any constant :

homogeneity

The -th cumulant is homogeneous in degree , be any constant, then:

Additivity

Let and be stochastically independent random variables, then holds for

For independent random variables, the characteristic function is a product and thus the logarithm is a sum:

The following applies to the sum of stochastically independent random variables :

Peculiarity of the normal distribution

For a normal distribution with expected value and variance , the characteristic function is the same and thus the cumulants:

for .

All accumulants greater than the 2nd order disappear. This property characterizes the normal distribution.

You can show that

  • either all accumulators except the first two disappear
  • or an infinite number of non-vanishing accumulants exist.

In other words: The cumulant generating function cannot be a finite polynomial of degree greater than 2.

Cumulants and Moments

Cumulants as a function of the moments

Denote the nth moment of a random variable . Through can be represented as

Consequently, the cumulants can be expressed by the moments or as follows:

In general, the dependence of the cumulants on the moments can be described by the following recursion formula:

Alternatively, the kth cumulant can be represented from Faà di Bruno's formula using the Bell polynomials and the moments as

.

With the central moments , the formulas are usually shorter:

The first two cumulants are of particular importance: is the expected value and is the variance . From the fourth order, the cumulative and central moment no longer coincide.

Derivation of the first cumulants

To develop to

and sets the series representation of

in the above development

If you sort according to powers of , you get the cumulants:

Moments as a function of the cumulants

The -th moment is a -th degree polynomial of the first cumulants. Here are the first six moments:

The coefficients correspond exactly to those in Faà di Bruno's formula . More generally, the -th moment is exactly the -th complete Bell polynomial , evaluated at the points :

.

In order to express the central moments as a function of the cumulants, neglect all terms in the above polynomials for the moments which appear as a factor.

Cumulants and volume partitions

Above we expressed the moments as polynomials in the cumulants. These polynomials have an interesting combinatorial interpretation: their coefficients count set partitions . The general form of these polynomials can be written as follows

in which

  • traverses the set of all partitions of an n-element set;
  • " " means that is one of the blocks into which the set was divided; and
  • is the size of the block .

Multivariate cumulants

The multivariate (or common ) cumulants of several random variables X 1 , ...,  X n can also be defined by a cumulant-generating function:

This formula can again be interpreted in combinatorial form according to

where runs through all partitions of {1, ...,  n  }, runs through the set of all blocks in the partition , and is the number of blocks in . For example we have

This combinatorial relationship between cumulants and moments takes on a simpler form if one expresses moments in terms of cumulants:

For example we then have:

The first cumulant of a random variable is its expected value, the common second cumulant of two random variables is their covariance . If some of the random variables are independent of each other, then any mixed cumulant which contains at least two of the independent variables vanishes. If all random variables are equal, the common cumulant is reduced to the usual nth cumulant of .

Another important property of the multivariate cumulants is multilinearity in the variables:

Inferences

The identically distributed and stochastically independent random variables are given .

Central limit theorem

For the random variable

Using the properties of homogeneity and additivity, the following cumulants result:

The order arises because the sum over the individual accumulants is of the order . Here are the orders of the first cumulants:

For is the order to the power of a negative exponent and thus the limit of infinitely many random variables applies:

I.e. only the first two cumulants remain. The only distribution that has only the first and second cumulants is the Gaussian distribution. This makes it plausible that the sum of any random variables divided by the square root of the number converges to the Gaussian distribution; this is the Central Limit Theorem . In order to complete this plausibility analysis to a proof, it is necessary to use general laws of characteristic functions . The Gaussian distribution thus has a special position among all distributions. If there are many stochastically independent influences in an experiment, then the totality of the influences can be represented by a Gaussian random variable.

As a simple special case, consider all random variables to be identical to mean value 0, variance and any higher moments.

For the random variable

one can use the displacement invariance of the cumulants of the order greater than or equal to 2. The only difference to the random variable is that the expected value is zero, even if the expected values ​​of the do not disappear.

Law of Large Numbers

For the random variable

Using the properties of homogeneity and additivity, the following cumulants result:

The order arises because the sum over the individual accumulants is of the order . Here are the orders of the first cumulants:

For is the order to the power of a negative exponent and thus the limit of infinitely many random variables applies:

I.e. only the first cumulative or the first moment remains. With increasing one obtains a Gaussian distribution around the mean value

,

where the latitude is of the order , and in the borderline case a sharp ( delta- shaped) peak at .

As a simple special case, consider all random variables to be identical to the mean , variance and any higher moments.

Thus is a random variable with the same mean as (called fair-expectation estimator for the mean of ). The width of the Gaussian distribution (standard deviation from the mean), which becomes ever narrower, is .

history

Cumulants and their properties were first described in 1889 by the Danish mathematician Thorvald Nicolai Thiele in a book published in Danish. Although this book was reported in detail in the yearbook on the progress of mathematics in the same year , the results were initially largely ignored, so that in 1901 Felix Hausdorff described these parameters as "newly introduced" (by him).

Free accumulators

In the above combinatorial moment cumulant formula

one adds up over all partitions of the set . If, instead, you only sum over non-crossing partitions , you get the free cumulants . These were introduced by Roland Speicher and play an analogous role in free probability theory as the usual cumulants in classical probability theory. In particular, the free cumulants are additive for free random variables. The Wigner's semicircular distribution , which is the free counterpart to the normal distribution, is characterized by the fact that only the free second order cumulant does not vanish.

literature

Individual evidence

  1. Thorvald Nicolai Thiele: Forelæsninger over almindelig Iagttagelseslære: Sandsynlighedsregning og least Kvadraters method , Copenhagen 1889.
  2. ^ Yearbook on the Progress of Mathematics JFM 21.0210.01 .
  3. Felix Hausdorff: Collected Works, Volume V: Astronomy, Optics and Probability Theory . 2006, ISBN 978-3-540-30624-5 , pp. 544, 577, doi: 10.1007 / 3-540-30669-2_8 .
  4. Speicher, Roland (1994), "Multiplicative functions on the lattice of non-crossing partitions and free convolution", Mathematische Annalen, 298 (4): 611-628
  5. Jonathan Novak, Piotr Śniady: What Is a Free Cumulant? . In: Notices of the American Mathematical Society . 58, No. 2, 2011, ISSN  0002-9920 , pp. 300-301.