147 lines
5.9 KiB
TeX
147 lines
5.9 KiB
TeX
|
\subsubsection{Application: Percolation}
|
||
|
|
||
|
|
||
|
We will now discuss another application of Kolmogorov's $0-1$-law, percolation.
|
||
|
|
||
|
\begin{definition}[\vocab{Percolation}]
|
||
|
Consider the graph with nodes $\Z^d$, $d \ge 2$, where edges from the lattice are added with probability $p$. The added edges are called \vocab[Percolation!Edge!open]{open};
|
||
|
all other edges are called
|
||
|
\vocab[Percolation!Edge!closed]{closed}.
|
||
|
|
||
|
More formally, we consider
|
||
|
\begin{itemize}
|
||
|
\item $\Omega = \{0,1\}^{\bE_d}$, where $\bE_d$ are all edges in $\Z^d$,
|
||
|
\item $\cF \coloneqq \text{product $\sigma$-algebra}$,
|
||
|
\item $\bP \coloneqq \left(p \underbrace{\delta_{\{1\} }}_{\text{edge is open}} + (1-p) \underbrace{\delta_{\{0\} }}_{\text{edge is absent closed}}\right)^{\otimes \bE_d}$.
|
||
|
\end{itemize}
|
||
|
\end{definition}
|
||
|
\begin{question}
|
||
|
Starting at the origin, what is the probability, that there exists
|
||
|
an infinite path (without moving backwards)?
|
||
|
\end{question}
|
||
|
\begin{definition}
|
||
|
An \vocab{infinite path} consists of an infinite sequence of distinct points
|
||
|
$x_0, x_1, \ldots$
|
||
|
such that $x_n$ is connected to $x_{n+1}$, i.e.~the edge $\{x_n, x_{n+1}\}$ is open.
|
||
|
\end{definition}
|
||
|
|
||
|
Let $C_\infty \coloneqq \{\omega | \text{an infinite path exists}\}$.
|
||
|
|
||
|
\begin{exercise}
|
||
|
Show that changing the presence / absence of finitely many edges
|
||
|
does not change the existence of an infinite path.
|
||
|
Therefore $C_\infty$ is an element of the tail $\sigma$-algebra.
|
||
|
Hence $\bP(C_\infty) \in \{0,1\}$.
|
||
|
\end{exercise}
|
||
|
Obviously, $\bP(C_\infty)$ is monotonic with respect to $p$.
|
||
|
For $d = 2$ it is known that $p = \frac{1}{2}$ is the critical value.
|
||
|
For $d > 2$ this is unknown.
|
||
|
|
||
|
% TODO: more in the notes
|
||
|
We'll get back to percolation later.
|
||
|
|
||
|
|
||
|
\section{Characteristic functions, weak convergence and the central limit theorem}
|
||
|
|
||
|
Characteristic functions are also known as the \vocab{Fourier transform}.
|
||
|
Weak convergence is also known as \vocab{convergence in distribution} / \vocab{convergence in law}.
|
||
|
We will abbreviate the central limit theorem by \vocab{CLT}.
|
||
|
|
||
|
So far we have dealt with the average behaviour,
|
||
|
\[
|
||
|
\frac{\overbrace{X_1 + \ldots + X_n}^{\text{i.i.d.}}}{n} \to \bE(X_1).
|
||
|
\]
|
||
|
We now want to understand \vocab{fluctuations} from the average behaviour,
|
||
|
i.e.\[
|
||
|
X_1 + \ldots + X_n - n \cdot \bE(X_1).
|
||
|
\]
|
||
|
% TODO improve
|
||
|
The question is, what happens on other timescales than $n$?
|
||
|
An example is
|
||
|
\[
|
||
|
\frac{X_1 + \ldots + X_n - n \bE(X_1)}{\sqrt{n} } \xrightarrow{n \to \infty} hv \cN(0, \Var(X_i)) (\ast)
|
||
|
\]
|
||
|
Why is $\sqrt{n}$ the right order? (Handwavey argument)
|
||
|
|
||
|
Suppose $X_1, X_2,\ldots$ are i.i.d. $\cN(0,1)$.
|
||
|
The mean of the l.h.s.~is $0$ and for the variance we get
|
||
|
\[
|
||
|
\Var(\frac{X_1 + \ldots + X_n - n \bE(X_1)}{\sqrt{n} }) = \Var\left( \frac{X_1+ \ldots + X_n}{\sqrt{n} } \right) = \frac{1}{n} \left( \Var(X_1) + \ldots + \Var(X_n) \right) = 1
|
||
|
\]
|
||
|
For the r.h.s.~we get a mean of $0$ and a variance of $1$.
|
||
|
So, to determine what $(\ast)$ could mean, it is necessary that $\sqrt{n}$
|
||
|
is the right scaling.
|
||
|
To define $(\ast)$ we need another notion of convergence.
|
||
|
This will be the weakest notion of convergence, hence it is called
|
||
|
\vocab{weak convergence}.
|
||
|
This notion of convergence will be defined in terms of characteristic functions of Fourier transforms.
|
||
|
|
||
|
\subsection{Characteristic functions and Fourier transform}
|
||
|
|
||
|
Consider $(\R, \cB(\R), \bP)$.
|
||
|
For every $t \in \R$ define a function $\phi(t) \coloneqq \phi_\bP(t) \coloneqq \int_{\R} e^{\i t x} \bP(dx)$.
|
||
|
We have
|
||
|
\[
|
||
|
\phi(t) = \int_{\R} \cos(tx) \bP(dx) + \i \int_{\R} \sin(tx) \bP(dx).
|
||
|
\]
|
||
|
\begin{itemize}
|
||
|
\item Since $|e^{\i t x}| \le 1$ the function $\phi(\cdot )$ is always defined.
|
||
|
\item We have $\phi(0) = 1$.
|
||
|
\item $|\phi(t)| \le \int_{\R} |e^{\i t x} | \bP(dx) = 1$.
|
||
|
\end{itemize}
|
||
|
We call $\phi_{\bP}$ the \vocab{characteristic function} of $\bP$.
|
||
|
|
||
|
\begin{remark}
|
||
|
Suppose $(\Omega, \cF, \bP)$ is an arbitrary probability space and
|
||
|
$X: (\Omega, \cF) \to (\R, \cB(\R))$ is a random variable.
|
||
|
Then we can define
|
||
|
\[
|
||
|
\phi_X(t) \coloneqq \bE[e^{\i t x}] = \int e^{\i t X(\omega)} \bP(d \omega) = \int_{\R} e^{\i t x} \mu(dx) = \phi_\mu(t)
|
||
|
\]
|
||
|
where $\mu = \bP x^{-1}$.
|
||
|
\end{remark}
|
||
|
|
||
|
\begin{theorem}[Inversion formula] % thm1
|
||
|
\label{inversionformula}
|
||
|
Let $(\Omega, \cB(\R), \bP)$ be a probability space.
|
||
|
Let $F$ be the distribution function of $\bP$
|
||
|
(i.e.~$F(x) = \bP((-\infty, x])$ for all $x \in \R$ ).
|
||
|
Then for every $a < b$ we have
|
||
|
\begin{eqnarray}
|
||
|
\frac{F(b) + F(b-)}{2} - \frac{F(a) + F(a-)}{2} = \lim_{T \to \infty} \frac{1}{2 \pi} \int_{-T}^T \frac{e^{-\i t b} - e^{- \i t a}}{- \i t} \phi(t) dt
|
||
|
\label{invf}
|
||
|
\end{eqnarray}
|
||
|
where $F(b-)$ is the left limit.
|
||
|
\end{theorem}
|
||
|
% TODO!
|
||
|
We will prove this later.
|
||
|
|
||
|
\begin{theorem}[Uniqueness theorem] % thm2
|
||
|
\label{charfuncuniqueness}
|
||
|
Let $\bP$ and $\Q$ be two probability measures on $(\R, \cB(\R))$.
|
||
|
Then $\phi_\bP = \phi_\Q \implies \bP = \Q$.
|
||
|
|
||
|
Therefore, probability measures are uniquely determined by their characteristic functions.
|
||
|
Moreover, \eqref{invf} gives a representation of $\bP$ (via $F$)
|
||
|
from $\phi$.
|
||
|
\end{theorem}
|
||
|
\begin{refproof}{charfuncuniqueness}
|
||
|
Assume that we have already shown \autoref{inversionformula}.
|
||
|
Suppose that $F$ and $G$ are the distribution functions of $\bP$ and $\Q$.
|
||
|
Let $a,b \in \R$ with $a < b$.
|
||
|
Assume that $a $ and $b$ are continuity points of both $F$ and $G$.
|
||
|
By \autoref{inversionformula} we have
|
||
|
\begin{IEEEeqnarray*}{rCl}
|
||
|
F(b) - F(a) = G(b) - G(a) \label{eq:charfuncuniquefg}
|
||
|
\end{IEEEeqnarray*}
|
||
|
|
||
|
Since $F$ and $G$ are monotonic, \autoref{eq:charfuncuniquefg}
|
||
|
holds for all $a < b$ outside a countable set.
|
||
|
|
||
|
Take $a_n$ outside this countable set, such that $a_n \ssearrow -\infty$.
|
||
|
Then, \autoref{eq:charfuncuniquefg} implies that
|
||
|
$F(b) - F(a_n) = G(b) - G(a_n)$ hence $F(b) = G(b)$.
|
||
|
Since $F$ and $G$ are right-continuous, it follows that $F = G$.
|
||
|
\end{refproof}
|
||
|
|