Distribution of a random variable

from Wikipedia, the free encyclopedia

The distribution of a random variable is a term from probability theory , a branch of mathematics. The distribution of a random variable makes it possible to extract information from a stochastic model that is “too large” and to assign meaningful probabilities to it again. An example of this is a lottery drawing: When modeling, the probabilities for each individual number combination are first defined. In general, however, one is not interested in the probability of drawing exactly a certain number sequence, but in how great the probability is for "n correct". To do this, a random variable is defined which extracts the information “number of correct persons”. The distribution of these random variables then indicates the probability that you have drawn “n correct”.

definition

Given a random variable from the probability space into the event space . Then it's called through

defined mapping the distribution of the random variables below . It defines a probability measure on . Here the archetype of below , i.e. the event . Sometimes there is also written for.

Examples

We consider a three-time coin toss as a model, modeled by the probability space with result set

,

Event system

and as a measure of probability, the equal distribution, since the coin is assumed to be fair and the tosses take place independently of one another , i.e.

.

The second event space is now defined as

,

the random variable counts the successes, so

In order to determine the distribution of these random variables, it is sufficient to go through a generator, i.e. here the individual elementary events. All other probabilities are then obtained by adding the probabilities of the (disjoint) producers. It is then

.

This is then the distribution of the random variables and a new probability measure on the event space .

Dimension theoretical perspective

From the point of view of measure theory , the distribution of a random variable is an image measure . The probability space corresponds to a special measurement space , the event space is identical to a measurement space and the random variable is a measurable function . Just like the image measure, the distribution of a random variable enables the “relocation” and modification of a probability measure from a measurement space to a measurement space.

Probability measures as distributions

General measures of probability as distributions

If a probability space is given, the probability measure can be represented as the distribution of a random variable in the following way: The event space is duplicated and the identical mapping of after is selected as the random variable . Then the probability measure and the distribution of the random variables agree. This justifies, among other things, the common term “probability distribution” for probability measures.

Real probability measures as distributions on a common probability space

In fact, each probability measure on the event space can be represented as a distribution of a random variable on the probability space. Here, the constant uniform distribution on the interval from 0 to 1 is used to describe the fact that each probability measure is clearly defined by its distribution function . If now is a distribution function of , then the quantile function defined by is chosen as the random variable

,

where is. This random variable now has the probability measure as a distribution. This statement makes it possible, for example, to examine any random variable in the real numbers for stochastic independence , since they can always be understood as random variables in the same probability space.

Probability measures constructed as distributions

Binomial distribution

The binomial distribution can be fundamentally defined as the distribution of a random variable. To do this, you define the simple coin toss of an unfair coin with the probability space , the number of successes, the amount of events and the probability measure . The n-times independent tossing of the coin is then described by the product model . Now we define a random variable of the product model to by

,

This is how this random variable models the number of successes in coin flips. The distribution of the random variables is then the binomial distribution, so .

Geometric distribution and negative binomial distribution

Just like the binomial distribution, the geometric distribution and the negative binomial distribution can be derived from a product model of a coin toss as the distribution of a random variable. In this case, the product model is the infinitely often repeated coin toss, i.e. with the same designations as above . The random variable from the product model in the event space defined by

then models the waiting time until the first success and has the geometric distribution as the distribution. If you model the waiting time for the nth success, you get the negative binomial distribution.

Generalizations

There are several special cases of the distribution of a random variable. The common distribution of random variables uses several random variables to define a multivariate distribution on a higher-dimensional space. The marginal distribution, on the other hand, is the distribution of a multivariate distribution under a coordinate mapping, it thus reduces the dimensionality of the probability distribution.

A variation of the distribution of a random variable is the conditional distribution and the regular conditional distribution . Both model additional prior knowledge about the outcome of the random experiment. The conditional distribution is easier to handle and is defined by the conditional probability , but has deficits in dealing with zero sets and is not as general. The regular conditional distribution requires the technical concept of the conditional expected value .

literature

Web links

Wiktionary: probability distribution  - explanations of meanings, word origins, synonyms, translations