Kolmogorov 0-1 Law

lean4-proofprobabilityvisualization
T tail eventP(T){0,1}\mathcal{T} \text{ tail event} \Rightarrow P(\mathcal{T}) \in \{0, 1\}

Statement

Let X1,X2,X_1, X_2, \ldots be independent random variables on a probability space (Ω,F,P)(\Omega, \mathcal{F}, P). The tail σ\sigma-algebra is

T=n=1σ(Xn,Xn+1,)=lim supnσ(Xn).\mathcal{T} = \bigcap_{n=1}^\infty \sigma(X_n, X_{n+1}, \ldots) = \limsup_{n\to\infty} \sigma(X_n).

A set ATA \in \mathcal{T} is called a tail event: it is not affected by changing any finite initial segment of the sequence.

Kolmogorov's 0-1 Law. For any tail event ATA \in \mathcal{T}:

P(A)=0orP(A)=1.P(A) = 0 \quad \text{or} \quad P(A) = 1.

Equivalently, any T\mathcal{T}-measurable random variable is PP-almost surely constant.

Visualization

Consider the event A={partial sums Sn are unbounded}A = \{\text{partial sums } S_n \text{ are unbounded}\} where Sn=X1++XnS_n = X_1 + \cdots + X_n and XiBernoulli(1/2)X_i \sim \text{Bernoulli}(1/2) (fair coin, ±1\pm 1). Observe:

  • Knowing X1,,X100X_1, \ldots, X_{100} does not determine whether supnSn=\sup_n S_n = \infty — this is a tail property.
  • Therefore ATA \in \mathcal{T}, and the law says P(A){0,1}P(A) \in \{0,1\}.
  • In this case the random walk is recurrent, so P(A)=1P(A) = 1.
Tail event structure:

σ(X₁,X₂,…) ⊇ σ(X₂,X₃,…) ⊇ σ(X₃,X₄,…) ⊇ … ⊇ T

T = intersection of all these = events independent of any finite prefix

Example tail events:
  {lim sup Sₙ/n = 0}   → probability 1  (by SLLN)
  {Sₙ bounded}          → probability 0  (recurrence)
  {∑Xₙ converges}       → probability 0 or 1 (depends on distribution)
  {Xₙ → 0}             → probability 0 or 1

Example non-tail event:
  {X₁ = 1}             → probability 1/2  (involves only X₁)

Proof Sketch

  1. Tail event is independent of each finite block. For fixed mm, the set Aσ(Xm+1,Xm+2,)A \in \sigma(X_{m+1}, X_{m+2}, \ldots) and σ(X1,,Xm)\sigma(X_1, \ldots, X_m) are independent (by independence of the XiX_i).
  2. Tail is independent of all finite blocks, hence of σ(mσ(X1,,Xm))\sigma(\bigcup_m \sigma(X_1,\ldots,X_m)). By a π\pi-λ\lambda argument (Dynkin's theorem), independence extends from π\pi-systems to the generated σ\sigma-algebras.
  3. The tail is independent of itself. Since ATσ(Xm+1,Xm+2,)A \in \mathcal{T} \subseteq \sigma(X_{m+1}, X_{m+2}, \ldots) for every mm, it is independent of σ(X1,,Xm)\sigma(X_1, \ldots, X_m) for all mm, hence of T\mathcal{T} itself.
  4. Self-independent implies 0-1. P(A)=P(AA)=P(A)2P(A) = P(A \cap A) = P(A)^2 forces P(A){0,1}P(A) \in \{0,1\}.

Connections

The law gives a zero-or-one dichotomy analogous to the determinism of the Strong Law of Large NumbersStrong Law of Large Numbers1ni=1nXia.s.E[X1]\frac{1}{n}\sum_{i=1}^n X_i \xrightarrow{\text{a.s.}} \mathbb{E}[X_1]The sample mean of i.i.d. integrable random variables converges almost surely to the population meanRead more → (the limiting frequency is PP-a.s. constant). The π\pi-λ\lambda argument used in step 2 is the same device behind the Monotone Convergence TheoremMonotone Convergence Theoremf monotone,  supnf(n)<    f(n)supnf(n)f \text{ monotone},\; \sup_n f(n) < \infty \implies f(n) \to \sup_n f(n)A bounded monotone sequence of reals always converges — the supremum is the limitRead more → in measure theory.

Lean4 Proof

import Mathlib.Probability.Independence.ZeroOne

open ProbabilityTheory MeasureTheory MeasurableSpace Filter

/-- Kolmogorov's 0-1 law: a tail-measurable set in an independent sequence
    has measure 0 or 1.
    Mathlib: `ProbabilityTheory.measure_zero_or_one_of_measurableSet_limsup_atTop`. -/
theorem kolmogorov_zero_one
    {Ω : Type*} {m0 : MeasurableSpace Ω}
    {μ : MeasureTheory.Measure Ω} [IsFiniteMeasure μ]
    {s :   MeasurableSpace Ω}
    (h_le :  n, s n  m0)
    (h_indep : iIndep s μ)
    {t : Set Ω}
    (ht : MeasurableSet[limsup s atTop] t) :
    μ t = 0  μ t = 1 :=
  measure_zero_or_one_of_measurableSet_limsup_atTop h_le h_indep ht