In probability theory , the Hoeffding inequality (after Wassilij Hoeffding ) describes the maximum probability that a sum of independent and limited random variables will deviate more than a constant from their expected value .
The Hoeffding inequality is also called the additive Chernoff inequality and is a special case of the Bernstein inequality .
sentence
Be independent random variables so that it almost certainly holds true. Also be a positive, real-valued constant. Then:
X
1
,
X
2
,
...
,
X
n
{\ displaystyle X_ {1}, X_ {2}, \ ldots, X_ {n}}
a
i
≤
X
i
-
E.
[
X
i
]
≤
b
i
{\ displaystyle a_ {i} \ leq X_ {i} - {\ textrm {E}} [X_ {i}] \ leq b_ {i}}
c
{\ displaystyle c}
Pr
[
∑
(
X
i
-
E.
[
X
i
]
)
≥
c
]
≤
exp
(
-
2
c
2
∑
(
b
i
-
a
i
)
2
)
.
{\ displaystyle \ Pr \ left [\ sum (X_ {i} - {\ textrm {E}} [X_ {i}]) \ geq c \ right] \ leq {\ textrm {exp}} \ left ({\ frac {-2c ^ {2}} {\ sum (b_ {i} -a_ {i}) ^ {2}}} \ right).}
proof
This proof follows the presentation by D. Pollard, see also Lutz Dümbgen's script (see literature).
To simplify the notation, consider the random variables with and furthermore, for an initially arbitrary map, the mapping that grows monotonically on the real numbers
. According to the Chebyshev inequality :
Y
i
=
X
i
-
E.
[
X
i
]
{\ displaystyle Y_ {i} = X_ {i} - {\ textrm {E}} [X_ {i}]}
E.
[
Y
i
]
=
0
{\ displaystyle {\ textrm {E}} [Y_ {i}] = 0}
z
>
0
{\ displaystyle z> 0}
x
↦
exp
(
z
x
)
{\ displaystyle x \ mapsto \ exp (zx)}
Pr
[
∑
Y
i
≥
c
]
≤
E.
[
exp
(
z
∑
Y
i
)
]
exp
(
z
c
)
=
exp
(
-
z
c
)
⋅
∏
E.
[
exp
(
z
Y
i
)
]
.
{\ displaystyle \ Pr \ left [\ sum Y_ {i} \ geq c \ right] \ leq {\ frac {{\ textrm {E}} [\ exp \ left (z \ sum Y_ {i} \ right)] } {\ exp (zc)}} = \ exp (-zc) \ cdot \ prod {\ textrm {E}} \ left [\ exp (zY_ {i}) \ right].}
Because of the convexity of the exponential function
exp
(
z
Y
i
)
=
exp
(
b
i
-
Y
i
b
i
-
a
i
z
a
i
+
Y
i
-
a
i
b
i
-
a
i
z
b
i
)
≤
b
i
-
Y
i
b
i
-
a
i
exp
(
z
a
i
)
+
Y
i
-
a
i
b
i
-
a
i
exp
(
z
b
i
)
,
{\ displaystyle \ exp (zY_ {i}) = \ exp \ left ({\ frac {b_ {i} -Y_ {i}} {b_ {i} -a_ {i}}} za_ {i} + {\ frac {Y_ {i} -a_ {i}} {b_ {i} -a_ {i}}} zb_ {i} \ right) \ leq {\ frac {b_ {i} -Y_ {i}} {b_ { i} -a_ {i}}} \ exp (za_ {i}) + {\ frac {Y_ {i} -a_ {i}} {b_ {i} -a_ {i}}} \ exp (zb_ {i }),}
and with it follows that
E.
[
Y
i
]
=
0
{\ displaystyle {\ textrm {E}} [Y_ {i}] = 0}
E.
[
exp
(
z
Y
i
)
]
≤
b
i
b
i
-
a
i
exp
(
z
a
i
)
+
-
a
i
b
i
-
a
i
exp
(
z
b
i
)
=
e
-
u
i
λ
i
(
(
1
-
λ
i
)
+
λ
i
e
u
i
)
{\ displaystyle {\ textrm {E}} \ left [\ exp (zY_ {i}) \ right] \ leq {\ frac {b_ {i}} {b_ {i} -a_ {i}}} \ exp ( za_ {i}) + {\ frac {-a_ {i}} {b_ {i} -a_ {i}}} \ exp (eg_ {i}) = e ^ {- u_ {i} \ lambda _ {i }} \ left (\ left (1- \ lambda _ {i} \ right) + \ lambda _ {i} e ^ {u_ {i}} \ right)}
for the constants and . Looking at the logarithm of the right hand side of this term
λ
i
=
-
a
i
b
i
-
a
i
{\ displaystyle \ lambda _ {i} = {\ frac {-a_ {i}} {b_ {i} -a_ {i}}}}
u
i
=
z
(
b
i
-
a
i
)
{\ displaystyle u_ {i} = z (b_ {i} -a_ {i})}
L.
(
u
,
λ
)
=
-
u
λ
+
log
(
(
1
-
λ
)
+
λ
e
u
)
,
{\ displaystyle L (u, \ lambda) = - u \ lambda + \ log \ left (\ left (1- \ lambda \ right) + \ lambda e ^ {u} \ right),}
so one can show by means of curve discussion and Taylor series expansion that it always
applies. If one reinserts this value into the first inequality due to the monotony of the exponential function as the upper bound, one obtains
L.
(
u
,
λ
)
≤
u
2
8th
{\ displaystyle L (u, \ lambda) \ leq {\ frac {u ^ {2}} {8}}}
Pr
[
∑
Y
i
≥
c
]
≤
exp
(
-
z
c
)
⋅
∏
exp
(
u
i
2
8th
)
=
exp
(
-
z
c
+
z
2
8th
∑
(
b
i
-
a
i
)
2
)
,
{\ displaystyle \ Pr \ left [\ sum Y_ {i} \ geq c \ right] \ leq \ exp (-zc) \ cdot \ prod \ exp ({\ frac {u_ {i} ^ {2}} {8 }}) = \ exp \ left (-zc + {\ frac {z ^ {2}} {8}} \ sum (b_ {i} -a_ {i}) ^ {2} \ right),}
which leads to the assertion to be proven if you choose .
z
=
4th
c
∑
(
b
i
-
a
i
)
2
{\ displaystyle z = {\ tfrac {4c} {\ sum (b_ {i} -a_ {i}) ^ {2}}}}
Examples
Consider the following question: how likely is it to total at least 500 if you roll the dice 100 times? Describes the outcome of the dice roll with , so , it follows the Hoeffding's inequality:
X
{\ displaystyle X}
E.
[
X
]
=
3
,
5
{\ displaystyle {\ textrm {E}} [X] = 3 {,} 5}
-
2
,
5
≤
X
-
E.
[
X
]
≤
2
,
5
{\ displaystyle -2 {,} 5 \ leq X - {\ textrm {E}} [X] \ leq 2 {,} 5}
Pr
[
∑
X
≥
500
]
=
Pr
[
∑
(
X
-
E.
[
X
]
)
≥
150
]
{\ displaystyle \ Pr \ left [\ sum X \ geq 500 \ right] = \ Pr \ left [\ sum (X - {\ textrm {E}} [X]) \ geq 150 \ right]}
≤
exp
(
-
2
⋅
150
2
∑
(
2
,
5
+
2
,
5
)
2
)
=
exp
(
-
45000
100
⋅
25th
)
=
exp
(
-
18th
)
≈
1.523
⋅
10
-
8th
{\ displaystyle \ leq \ exp \ left ({\ frac {-2 \ cdot 150 ^ {2}} {\ sum (2 {,} 5 + 2 {,} 5) ^ {2}}} \ right) = \ exp \ left ({\ frac {-45000} {100 \ cdot 25}} \ right) = \ exp \ left (-18 \ right) \ approx 1 {,} 523 \ cdot 10 ^ {- 8}}
literature
Wassily Hoeffding, "Probability Inequalities for Sums of Bounded Random Variables," Journal of the American Statistical Association, Vol. 58, 1963, pp. 13-30.
David Pollard, "Convergence of Stochastic Processes", Springer Verlag, 1984.
Lutz Dümbgen, Empirical Processes , University of Bern, 2010.
Otto Kerner, Joseph Maurer, Jutta Steffens, Thomas Thode and Rudolf Voller, Vieweg Mathematik Lexikon , second revised edition, Vieweg Verlag, 1993.
<img src="https://de.wikipedia.org/wiki/Special:CentralAutoLogin/start?type=1x1" alt="" title="" width="1" height="1" style="border: none; position: absolute;">