Probability theory
The probability theory , and probability theory or probabilistic , is a branch of mathematics that from the formalization, modeling and analysis of random events emerged. Together with mathematical statistics , which make statements about the underlying model based on observations of random processes, they form the mathematical subarea of stochastics .
The central objects of probability theory are random events , random variables and stochastic processes .
Axiomatic structure
Like every branch of modern mathematics, probability theory is formulated in set theory and built on axiomatic specifications. The starting point of probability theory are events that are understood as sets and to which probabilities are assigned; Probabilities are real numbers between 0 and 1; the assignment of probabilities to events must meet certain minimum requirements.
These definitions give no indication of how to determine the probabilities of individual events; they also say nothing about what chance and what probability actually are. The mathematical formulation of probability theory is thus open to various interpretations, but its results are nevertheless exact and independent of the respective understanding of the concept of probability.
Definitions
Conceptually, the mathematical consideration is based on a random process or random experiment . All possible results of this random process are summarized in the result set . Often one is not interested in the exact result at all , but only in whether it is in a certain subset of the result set, which can be interpreted to mean that an event has occurred or not. So an event is defined as a subset of . If the event contains exactly one element of the result set, it is an elementary event . Compound events contain multiple outcomes. The result is therefore an element of the result set, but the event is a subset.
So you can assign probabilities to the events in a meaningful way, they are listed in a quantitative system, the algebra of events or the event system on , a set of subsets of , for which: It contains and is a σbody , d. That is, it is closed with respect to the set operations of union and complement formation (relative with respect to ) as well as with respect to the infinite union of countably many sets. The probabilities are then images of a certain mapping of the event space into the interval [0,1]. Such a mapping is called a probability measure . The triple is called the probability space.
Axioms of Kolmogorov
The axiomatic foundation of probability theory was developed by Andrei Kolmogorow in the 1930s . A probability measure must therefore satisfy the following three axioms:
Axioms:
 For each event, the probability of is a real number between 0 and 1: .
 The certain event has a probability of 1: .
 The probability of a union of countably many incompatible events is equal to the sum of the probabilities of the individual events. Events are called incompatible if they are disjoint in pairs, i.e. for all . It is therefore true . This property is also called σadditivity .
Example: As part of a physical model, a probability measure is used to describe the outcome of a coin toss, the possible outcomes ( called events ) may be numbers and heads .
 Then is the result set .
 The power set can be chosen as the event space, thus .
 For the measure of probability it is clear from the axioms:
Additional physical assumptions about the nature of the coin can now lead to a choice .
Inferences
From the axioms there are some direct consequences:
1. From the additivity of probability of disjoint events follows that complementary events (counterevents) complementary probabilities ( against probabilities ) have: .
 Proof: It is as well . Consequently, according to axiom (3): and then Ax (2): . Changed follows: .
2. It follows that the impossible event, the empty set , the probability zero has: .
 Proof: It is , and so on Axiom (3): . It follows from this .
3. For the union not necessary disjoint events follows: .
 Proof: The quantities required for the proof are shown in the picture above. The set can then be represented as the union of three disjoint sets:
 According to (3) it follows: .
 On the other hand, according to (3), both
 as well as
 .
 Addition gives:
 .
 Rearranging results .
 The PoincaréSylvester sieve formula generalizes this assertion in the case of n different (not necessarily disjoint) subsets.
Furthermore, a distinction must be made between countable and uncountable result sets.
Countable result set
With a countable result set, each elementary event can be assigned a positive probability. If is finite or countably infinite, one can choose the power set of for σalgebra . The sum of the probabilities of all natural events from is here 1.
Uncountable result set
A prototype of an uncountable result set is the set of real numbers. In many models it is not possible to meaningfully assign a probability to all subsets of the real numbers. As an event system, instead of the power set of the real numbers, one usually chooses the Borel σalgebra , that is the smallest σalgebra that contains all intervals of real numbers as elements. The elements of this σalgebra are called Borel sets or also ( Borel ) measurable. If the probability of any Borel set as an integral
can be written over a probability density is called absolutely continuous . In this case (but not only in this case) all elementary events { x } have the probability 0. The probability density of an absolutely continuous probability measure is only uniquely determined almost everywhere, i. . e, they can be applied to any Lebesgue  null set , so an amount of Lebesgue measure 0 will be modified without being changed. If the first derivative of the distribution function of exists, then it is a probability density of P. However, the values of the probability density are not interpreted as probabilities.
Special properties in the case of discrete probability spaces
Laplace experiments
If one assumes that only a finite number of natural events are possible and that all are equal, i. H. occur with the same probability (such as when tossing an ideal coin, where {tails} and {heads} each have a probability of 0.5), one speaks of a Laplace experiment . Then probabilities can be calculated easily: We assume a finite result set that has the cardinality , i.e. i.e., it has elements. Then the probability of each natural event is simple .
 Proof: If is, then there are natural events . It is then on the one hand and on the other hand two elementary events are disjoint (incompatible: if one occurs, the other cannot occur). So the conditions for axiom (3) are fulfilled, and we have:
 Since on the other hand it is supposed to be, and therefore rearranged: as claimed.
As a consequence it follows that for events that are composed of several elementary events, the corresponding multiple probability applies. If an event is powerful , it is the union of elementary events . Each of these has the probability , so is . So you get the simple connection
In Laplace's experiments, the probability of an event is equal to the number of outcomes that are favorable to that event divided by the total number of possible outcomes.
The following is an example of rolling the dice with an ideal dice.
 ⚀⚁⚂⚃⚄⚅
 ⚄⚅
The event = high number (5 or 6) has a probability of 1/3.
A typical attempt at Laplace is also drawing a card from a game of cards or drawing a ball from an urn with balls. Here every elementary event has the same probability. Combinatorial methods are often used to determine the number of elementary events in Laplace experiments .
The concept of the Laplace experiments can be generalized to the case of a constant uniform distribution .
Conditional probability
A conditional probability is understood as the probability of an event occurring , provided that another event is already known. Of course , it must be able to happen, so it cannot be the impossible event. Then or less often one writes for “probability of under the assumption ”, in short “ of , provided ”.
Example: The probability of drawing a heart card from a Skat sheet (event ) is 1/4, because there are 32 cards and 8 of them are heart cards. Then is . The counterevent is then diamonds, spades or clubs and therefore has the probability .
If, however, the event “The card is red” has already occurred (a heart or diamond card was drawn, but it is not known which of the two colors), you only have the choice between the 16 red cards , then the probability is that it is then the heart leaf.
This consideration applied to a Laplace experiment. For the general case , the conditional probability of “ provided ” is defined as
That this definition is meaningful is shown by the fact that the probability defined in this way satisfies Kolmogorov's axioms if one restricts oneself to a new result set; d. i.e. that the following applies:
 If pairs are disjoint, then
Proof:

is the quotient of two probabilities for which axiom (1) holds and . Since the impossible event is not supposed to be, even is . So also applies to the quotient . Furthermore, and are disjoint, and their union is . So by Axiom (3): . There is, follows and therefore .
 It is
 Furthermore:
 This was to be shown.
Example: Let it be as above the event “Drawing a heart card” and the event “It is a red card”. Then:
and
Hence:
The following consequences result from the definition of the conditional probability:
Association probability (intersections of events)
The simultaneous occurrence of two events and corresponds in set theory to the occurrence of the compound event . The likelihood thereof calculated for joint probability or joint probability
Proof: According to the definition of the conditional probability, on the one hand
and on the other hand too
Switching to then immediately delivers the assertion.
Example: A card is drawn from 32 cards. be the event: "It is a king". be the event: "It's a heart card". Then the simultaneous occurrence of and , thus the event: “The card drawn is a king of hearts”. Apparently it is . Furthermore , because there is only one heart card among the four kings. Indeed, then the probability is for the King of Hearts.
Bayes' theorem
The conditional probability of the condition can be from by the conditional probability under the condition by
express if one knows the total probabilities and ( Bayes' theorem ).
Dependence and independence from events
Events are called independent of one another if the occurrence of one does not affect the probability of the other. In the opposite case, it is called dependent. One defines:
 Two events and are independent if applies.
 Inaccurately but memorably worded: In the case of independent events, the probabilities can be multiplied.
That this does justice to the term "independence" can be seen by changing over to :
This means: The total probability for is just as great as the probability for , provided ; so the occurrence of does not affect the probability of .
Example: One of 32 cards is drawn. be the event "It's a heart card". be the event "It is a picture card". These events are independent, because the knowledge that you draw a picture card does not affect the probability that it is a heart card (the proportion of heart cards among the picture cards is just as large as the proportion of heart cards. Cards on all cards). Apparently is and . is the event "It is a heart picture card". Since there are three of them, is . And in fact you find that is.
Another example of very small and very large probabilities can be found in the Infinite Monkey Theorem .
Dimension theory perspective
Classical probability calculus only considers probabilities on discrete probability spaces and continuous models with density functions. These two approaches can be unified and generalized through the modern formulation of probability theory, which is based on the concepts and results of the theory of measure and integration .
Probability spaces
In this view, a probability space is a measure space with a probability measure . This means that the result set is any set, the event space is a σalgebra with a basic set and is a measure that is normalized by .
Important standard cases of probability spaces are:
 is a countable set and is the power set of . Then every probability uniquely defined by its values on the oneelement subsets of and for all true
 .
 is a subset of and is the Borel σalgebra on . If the probability measure is absolutely continuous with respect to the Lebesgue measure , then according to the RadonNikodým theorem it has a Lebesgue density , i.e. h., for all true
 .
 Conversely, for a nonnegative measurable function that fulfills the normalization condition, this formula defines a probability measure .
 is a Cartesian product and is the product σalgebra of σalgebras based on . If probability measures are given, then the product measure defines a probability measure which models the independent execution of the individual experiments one after the other.
Random variable
A random variable is the mathematical concept for a quantity whose value depends on chance. From maßtheoretischer point of view it is a measurable function on a probability space into a measuring space consisting of a set and a σalgebra on . Measurability means that the archetype is an element of σalgebra for everyone . The distribution of is then nothing other than the image size
 ,
which is induced by on the measurement space and makes it a probability space .
The expected value of a realvalued random variable averages the possible results. It can be defined abstractly as the integral of with respect to the probability measure :
 .
Probability Theory and Statistics
Probability theory and mathematical statistics are collectively referred to as stochastics . Both areas are closely interrelated:
 Statistical distributions are regularly modeled under the assumption that they are the result of random processes.
 Statistical methods can provide indications of the behavior of probability distributions in a numerical way.
application areas
The theory of probability arose from the problem of the fair distribution of the stakes in abandoned games of chance . Other early uses also came from the area of gambling.
Today probability theory is a foundation of statistics . Applied statistics use the results of probability theory to analyze survey results or to make economic forecasts.
Large areas of physics such as thermodynamics and quantum mechanics use probability theory for the theoretical description of their results.
It is also the basis for mathematical disciplines such as reliability theory, renewal theory and queuing theory and the tool for analysis in these areas.
Probability theory is also of central importance in pattern recognition .
Probability Theory in School
Due to its diverse areas of application and the everyday relevance of even young students, probability theory is taught from grade 1 in all types of school as part of mathematics lessons. While elementary school is still about getting to know the basic concepts of probability calculus and evaluating the first random experiments with regard to their chances of winning, in lower secondary school the concept of probability is increasingly being examined analytically in its diversity and increasingly complex random experiments are the focus of interest. In the upper secondary level, the previous knowledge is expanded to include specific aspects such as Bernoulli chains, conditional probability and Laplace experiments.
See also
Literature (selection)
 Robert B. Ash : Real Analysis and Probability (= Probability and Mathematical Statistics . Volume 11. ). Academic Press, Inc. , New York (et al.) 1972, ISBN 0120652013 . MR0474442
 Krishna B. Athreya , Soumendra N. Lahiri : Measure Theory and Probability Theory . Springer Verlag , New York 2006, ISBN 9780387329031 . MR2247694
 Heinz Bauer : Probability Theory and Basics of Measure Theory . 4th edition. de Gruyter , Berlin 1991, ISBN 3110121913 .
 Heinz Bauer: Probability Theory . 5th, revised and improved edition. de Gruyter, Berlin, New York 2002, ISBN 3110172364 . MR1902050
 Kai Lai Chung : A Course in Probability Theory . Academic Press, San Diego (et al.) 2001, ISBN 0121741516 . MR1796326
 Bruno de Finetti : Probability Theory . Introductory synthesis with a critical appendix. 4th edition. R. Oldenbourg Verlag , Munich (inter alia) 1981, ISBN 3486447017 . MR0742141
 Harald Cramér : Mathematical Methods of Statistics (= Princeton Mathematical Series ). 11th edition. Princeton University Press , Princeton 1966.
 Richard M. Dudley : Real Analysis and Probability (= Cambridge Studies in Advanced Mathematics . Volume 74 ). Cambridge University Press , Cambridge 2002, ISBN 0521007542 . MR1932358
 P. Gänssler , W. Stute : Probability Theory (= university text . Volume 91 ). Springer Verlag, Berlin Heidelberg, New York 1977, ISBN 3540084185 . MR0501219
 Marek Fisz : Probability calculation and mathematical statistics (= university books for mathematics . Volume 40 ). 8th edition. VEB Deutscher Verlag der Wissenschaften , Berlin 1976.
 Boris Vladimirovich Gnedenko : Textbook of the theory of probability . Verlag Harri Deutsch , Thun, Frankfurt am Main 1997, ISBN 3817115318 .
 HansOtto Georgii : Stochastics . Introduction to probability theory and statistics. 5th edition. de Gruyter, 2015, ISBN 9783110359695 .
 J. HoffmannJørgensen : Probability with a View toward Statistics . Volume I (= Chapman & Hall Probability Series . Volume 91 ). Chapman and Hall , New York 1994, ISBN 0412052210 . MR1278485
 Achim Klenke : Probability Theory . 3rd, revised and expanded edition. Springer Spectrum , Berlin, Heidelberg 2013, ISBN 9783642360176 , doi : 10.1007 / 9783642360186 .
 Oleg Klesov : Limit Theorems for MultiIndexed Sums of Random Variables . Springer Verlag, Heidelberg, New York, Dordrecht, London 2014, ISBN 9783662443873 , doi : 10.1007 / 9783662443880 . MR3244237
 A. Kolmogoroff : Basic concepts of probability theory . Reprint (= results of mathematics and its border areas . Volume 3 ). Springer Verlag, Berlin, Heidelberg, New York 1973, ISBN 354006110X . MR0494348
 A. Kolmogoroff: About the sums by chance of certain independent quantities . In: Mathematical Annals . tape 99 , 1928, pp. 309319 , doi : 10.1007 / BF01459098 . MR1512588
 AJ Khintchine and AN Kolmogoroff: On the convergence of series, the terms of which are determined by chance . In: Recueil mathématique de la Société mathématique de Moscou [Matematicheskii Sbornik] . tape 32 , 1925, pp. 668677 .
 Ulrich Krengel : Introduction to probability theory and statistics . For studies, professional practice and teaching (= Vieweg Studium: advanced course in mathematics ). 8th expanded edition. Vieweg , Wiesbaden 2005, ISBN 3834800635 .
 Norbert Kusolitsch : Measure and probability theory . An introduction (= Springer textbook ). 2nd, revised and expanded edition. SpringerVerlag, Berlin, Heidelberg 2014, ISBN 9783642453861 , doi : 10.1007 / 9783322964182 .
 RG Laha , VK Rohatgi : Probability Theory (= Wiley Series in Probability and Mathematical Statistics ). John Wiley & Sons , New York (et al.) 1979, ISBN 047103262X . MR0534143
 Michel Ledoux , Michel Talagrand : Probability in Banach Spaces . Isoperimetry and Processes (= results of mathematics and its border areas (3rd part) . Volume 23 ). Springer Verlag, Berlin (inter alia) 1991, ISBN 3540520139 . MR1102015
 Richard von Mises : Probability, Statistics and Truth . Reprint of the 1957 English edition. Dover Publications, Inc. , New York 1981, ISBN 0486242145 . MR0668875
 Jacques Neveu : Mathematical Foundations of Probability Theory . Translated from the French by Karl Bosch . R. Oldenbourg Verlag, Munich, Vienna 1969. MR0245056
 Alfréd Rényi : Probability Theory . With an appendix on information theory (= university books for mathematics . Volume 54 ). 5th edition. Deutscher Verlag der Wissenschaften, Berlin 1977. MR0474442
 AN Širjaev : Probability (= university books for mathematics . Volume 91 ). Deutscher Verlag der Wissenschaften, Berlin 1988, ISBN 3326001959 . MR0967761
 Vladimir Spokoiny , Thorsten Dickhaus : Basics of Modern Mathematical Statistics (= Springer Texts in Statistics ). SpringerVerlag, Heidelberg, New York, Dordrecht, London 2015, ISBN 9783642399084 . MR3289985
 JV Uspensky : Introduction to Mathematical Probability . MacGrawHill Book Company, Inc. , New York, London 1937.
 NN Vakhania , VI Tarieladze , SA Chobanyan : Probability Distributions on Banach Spaces (= Mathematics and its Applications (Soviet Series) . Volume 14 ). D. Reidel Publishing Company , Dordrecht, Boston, Lancaster, Tokyo 1987, ISBN 9027724962 .
 Walter Vogel : Probability Theory (= Studia Mathematica . Volume XXII ). Vandenhoeck & Ruprecht , Göttingen 1970. MR0286145
Web links
Individual evidence
 ↑ https://kultusministerium.hessen.de/schulsystem/bildungsstandardskerncurriculaundlehrplaene/kerncurricula/primarstufe/mathematik
 ↑ https://kultusministerium.hessen.de/schulsystem/bildungsstandardskerncurriculaundlehrplaene/kerncurricula/sekundarstufei/mathematik
 ↑ https://kultusministerium.hessen.de/schulsystem/bildungsstandardskerncurriculaundlehrplaene/kerncurricula/gymnasialeoberstufe12