Bernstein Inequality (Stochastics)

from Wikipedia, the free encyclopedia

The Bernstein inequality is an estimate from probability theory and was proven by Sergei Bernstein (1880–1968). It is an extension of the Hoeffding inequality , the estimation of which can be improved by adding an additional requirement to the variance of the random variables.

The inequality holds for arbitrarily distributed limited random variables with vanishing expectation and limited variance.

sentence

Be a probability space . Let and be positive real numbers and a natural number. be independently distributed random variables with and for all .

Then:

lemma

The following applies to all :

A proof of the lemma and a more detailed proof of the theorem are in the evidence archive .

Proof (proposition)

This proof follows the approach from "Support Vector Machines" (see literature).

To simplify, define

After switching, it results:

Now the inequality is shown. Apply the Markov inequality with and for a (t will be chosen specifically later):

Since the random variables are independent according to the assumption, the product and the expected value can be exchanged. Form an infinite exponential series from the exponential function. This is limited by the integrable majorante . So expectation and sum can be swapped. With and the prerequisite follows:

With the estimate it follows:

The inequality for gives with :

Put :

From the lemma (above) follows with .

application

Problem 1

Adopted and known.

What is the maximum probability that the mean value exceeds a given value k?

Dissolve after .

The probability that the mean value exceeds the value k is at most .

Problem 2

Would continue to be and known.

The following should apply:

Calculate k using Bernstein's inequality .

Problem 3

Be and known.

How many realizations are required at least, so that for given values and , that

 ?

You need at least realizations.

example

Consider a coin as a random variable. We denote the i th coin toss with . This applies to heads and tails .

What is the probability that the mean value will exceed the value after throws ?

Since the probability for heads and tails is 50% each, the expectation value is 0. Since the random variables can only take the values ​​−1 and 1, a choice can be made.

. So you can choose .

Now choose one more time . Where is the number of coin flips. According to the Bernstein inequality:

So, for example, with 50 throws:

For 50 throws to exceed the mean , you would have to throw your head 31 times.

That is, the probability of getting 31 heads in 50 tosses is less than

See also

literature

  • Ingo Steinwart, Andreas Christmann: Support Vector Machines . Information Science and Statistics. 1st edition. Springer, Berlin 2008

Web links