From HandWiki - Reading time: 6 min
| Part of a series on statistics |
| Probability theory |
|---|
The standard probability axioms are the foundations of probability theory introduced by Russian mathematician Andrey Kolmogorov in 1933.[1] Like all axiomatic systems, they outline the basic assumptions underlying the application of probability to fields such as pure mathematics and the physical sciences, while avoiding logical paradoxes.[2]
The probability axioms do not specify or assume any particular interpretation of probability, but may be motivated by starting from a philosophical definition of probability and arguing that the axioms are satisfied by this definition. For example,
The third axiom, σ-additivity, is relatively modern, and originates with Lebesgue's measure theory. Some authors replace this with the strictly weaker axiom of finite additivity, which is sufficient to deal with some applications.[5]
In order to state the Kolmogorov axioms, the following pieces of data must be specified:
Taken together, these assumptions mean that is a measure space. It is additionally assumed that , making this triple a probability space.[1]
The probability of an event is a non-negative real number. This assumption is implied by the fact that is a measure on .
Theories which assign negative probability relax the first axiom.
This is the assumption of unit measure: that the probability that at least one of the elementary events in the entire sample space will occur is 1.From this axiom it follows that is always finite, in contrast with more general measure theory.
This is the assumption of σ-additivity: Any countable sequence of disjoint sets (synonymous with mutually exclusive events) satisfies
This property again is implied by the fact that is a measure. Note that, by taking and for all , one deduces that . This in turn shows that σ-additivity implies finite additivity.
Some authors consider merely finitely additive probability spaces, in which case one just needs an algebra of sets, rather than a σ-algebra.[6] Quasiprobability distributions in general relax the third axiom.
In order to demonstrate that the theory generated by the Kolmogorov axioms corresponds with classical probability, some elementary consequences are typically derived.[7]
By dividing into the disjoint sets , and , one arrives at a probabilistic version of the inclusion-exclusion principle[9]In the case where is finite, the two identities are equivalent.
In order to actually do calculations when is an infinite set, it is sometimes useful to generalize from a finite sample space. For example, if consists of all infinite sequences of tosses of a fair coin, it is not obvious how to compute the probability of any particular set of sequences (i.e. an event). If the event is "every flip is heads", then it is intuitive that the probability can be computed as:In order to make this rigorous, one has to prove that is continuous, in the following sense. If is a sequence of events increasing (or decreasing) to another event , then[10]
Consider a single coin-toss, and assume that the coin will either land heads (H) or tails (T) (but not both). No assumption is made as to whether the coin is fair.[11]
We may define:
Kolmogorov's axioms imply that:
The probability of neither heads nor tails, is 0.
The probability of either heads or tails, is 1.
The sum of the probability of heads and the probability of tails, is 1.