2023-06-29 20:31:46 +02:00
|
|
|
\lecture{21}{2023-06-29}{}
|
2023-06-29 22:18:23 +02:00
|
|
|
\subsection{An Application of the Optional Stopping Theorem}
|
2023-06-29 20:31:46 +02:00
|
|
|
|
|
|
|
This is the last lecture relevant for the exam.
|
|
|
|
(Apart from lecture 22 which will be a repetion).
|
|
|
|
|
|
|
|
\begin{goal}
|
|
|
|
We want to see an application of the
|
2023-07-28 03:45:37 +02:00
|
|
|
\ref{optionalstopping}.
|
2023-06-29 20:31:46 +02:00
|
|
|
\end{goal}
|
|
|
|
|
|
|
|
\begin{notation}
|
|
|
|
Let $E$ be a complete, separable metric space (e.g.~$E = \R$).
|
|
|
|
Suppose that for all $x \in E$ we have a probability measure
|
2023-06-29 20:39:53 +02:00
|
|
|
$\mathbf{P}(x, \dif y)$ on $E$.
|
2023-06-29 20:31:46 +02:00
|
|
|
% i.e. $\mu(A) \coloneqq \int_A \bP(x, \dif y)$ is a probability measure.
|
|
|
|
Such a probability measure is a called
|
|
|
|
a \vocab{transition probability measure}.
|
|
|
|
\end{notation}
|
2023-06-29 20:39:53 +02:00
|
|
|
\begin{example}
|
2023-06-29 20:31:46 +02:00
|
|
|
$E =\R$,
|
2023-06-29 20:39:53 +02:00
|
|
|
\[\mathbf{P}(x, \dif y) = \frac{1}{\sqrt{2 \pi} } e^{- \frac{(x-y)^2}{2}} \dif y\]
|
2023-06-29 20:31:46 +02:00
|
|
|
is a transition probability measure.
|
2023-06-29 20:39:53 +02:00
|
|
|
\end{example}
|
2023-06-29 20:31:46 +02:00
|
|
|
\begin{example}[Simple random walk as a transition probability measure]
|
2023-06-29 20:39:53 +02:00
|
|
|
$E = \Z$, $\mathbf{P}(x, \dif y)$
|
2023-06-29 20:31:46 +02:00
|
|
|
assigns mass $\frac{1}{2}$ to $y = x+1$ and $y = x -1$.
|
|
|
|
\end{example}
|
|
|
|
|
|
|
|
\begin{definition}
|
|
|
|
For every bounded, measurable function $f : E \to \R$,
|
|
|
|
$x \in E$
|
|
|
|
define
|
|
|
|
\[
|
2023-06-29 20:39:53 +02:00
|
|
|
(\mathbf{P} f)(x) \coloneqq \int_E f(y) \mathbf{P}(x, \dif y).
|
2023-06-29 20:31:46 +02:00
|
|
|
\]
|
2023-06-29 20:39:53 +02:00
|
|
|
This $\mathbf{P}$ is called a \vocab{transition operator}.
|
2023-06-29 20:31:46 +02:00
|
|
|
\end{definition}
|
|
|
|
\begin{fact}
|
2023-06-29 20:39:53 +02:00
|
|
|
If $f \ge 0$, then $(\mathbf{P} f)(\cdot ) \ge 0$.
|
2023-06-29 20:31:46 +02:00
|
|
|
|
2023-06-29 20:39:53 +02:00
|
|
|
If $f \equiv 1$, we have $(\mathbf{P} f) \equiv 1$.
|
2023-06-29 20:31:46 +02:00
|
|
|
\end{fact}
|
|
|
|
|
|
|
|
\begin{notation}
|
2023-06-29 20:39:53 +02:00
|
|
|
Let $\mathbf{I}$ denote the \vocab{identity operator},
|
2023-06-29 20:31:46 +02:00
|
|
|
i.e.
|
|
|
|
\[
|
2023-06-29 20:39:53 +02:00
|
|
|
(\mathbf{I} f)(x) = f(x)
|
2023-06-29 20:31:46 +02:00
|
|
|
\]
|
|
|
|
for all $f$.
|
2023-06-29 20:39:53 +02:00
|
|
|
Then for a transition operator $\mathbf{P}$ we write
|
2023-06-29 20:31:46 +02:00
|
|
|
\[
|
2023-06-29 20:39:53 +02:00
|
|
|
\mathbf{L} \coloneqq \mathbf{I} - \mathbf{P}.
|
2023-06-29 20:31:46 +02:00
|
|
|
\]
|
|
|
|
\end{notation}
|
|
|
|
|
|
|
|
\begin{goal}
|
|
|
|
Take $E = \R$.
|
|
|
|
Suppose that $A^c \subseteq \R$ is a bounded domain.
|
|
|
|
Given a bounded function $f$ on $\R$,
|
|
|
|
we want a function $u$ which is bounded,
|
|
|
|
such that
|
2023-06-29 22:18:23 +02:00
|
|
|
$\mathbf{L}u = 0$ on $A^c$ and $u = f$ on $A$.
|
2023-06-29 20:31:46 +02:00
|
|
|
\end{goal}
|
|
|
|
|
|
|
|
We will show that $u(x) = \bE_x[f(X_{T_A})]$
|
|
|
|
is the unique solution to this problem.
|
|
|
|
|
|
|
|
\begin{definition}
|
|
|
|
Let $(\Omega, \cF, \{\cF_n\}_n, \bP_x)$
|
|
|
|
be a filtered probability space, where for every $x \in \R$,
|
|
|
|
$\bP_x$ is a probability measure.
|
2023-06-29 20:39:53 +02:00
|
|
|
Let $\bE_x$ denote expectation with respect to $\mathbf{P}(x, \cdot )$.
|
2023-06-29 20:31:46 +02:00
|
|
|
Then $(X_n)_{n \ge 0}$ is a \vocab{Markov chain} starting at $x \in \R$
|
|
|
|
with \vocab[Markov chain!Transition probability]{transition probability}
|
2023-06-29 20:39:53 +02:00
|
|
|
$\mathbf{P}(x, \cdot )$ if
|
2023-06-29 20:31:46 +02:00
|
|
|
\begin{enumerate}[(i)]
|
|
|
|
\item $\bP_x[X_0 = x] = 1$,
|
|
|
|
\item for all bounded, measurable $f: \R \to \R$,
|
|
|
|
\[\bE_x[f(X_{n+1}) | \cF_n] \overset{\text{a.s.}}{=}%
|
|
|
|
\bE_{x}[f(X_{n+1}) | X_n] = %
|
2023-06-29 20:39:53 +02:00
|
|
|
\int f(y) \mathbf{P}(X_n, \dif y).\]
|
2023-06-29 20:31:46 +02:00
|
|
|
\end{enumerate}
|
|
|
|
(Recall $\cF_n = \sigma(X_1,\ldots, X_n)$.)
|
|
|
|
\end{definition}
|
|
|
|
\begin{example}
|
|
|
|
Suppose $B \in \cB(\R)$ and $f = \One_B$.
|
|
|
|
Then the first equality of (ii) simplifies to
|
|
|
|
\[
|
|
|
|
\bP_x[X_{n+1} \in B | \cF_n] = \bP_x[X_{n+1} \in B | \sigma(X_n)].
|
|
|
|
\]
|
|
|
|
\end{example}
|
|
|
|
|
|
|
|
\begin{example}
|
|
|
|
Let $\xi_i$ be i.i.d.~with$\bP[\xi_i = 1] = \bP[\xi_i = -1] = \frac{1}{2}$
|
|
|
|
and define $X_n \coloneqq \sum_{i=1}^{n} \xi_i$.
|
|
|
|
|
|
|
|
Intuitively, conditioned on $X_n$, $X_{n+1}$ should
|
|
|
|
be independent of $\sigma(X_1,\ldots, X_{n-1})$.
|
|
|
|
|
2023-06-29 22:18:23 +02:00
|
|
|
\begin{claim*}
|
2023-06-29 20:39:53 +02:00
|
|
|
For a set $B$, we have
|
2023-06-29 22:18:23 +02:00
|
|
|
\[
|
|
|
|
\bE[\One_{X_{n+1} \in B} | \sigma(X_1,\ldots, X_n)] %
|
|
|
|
= \bE[\One_{X_{n+1} \in B} | \sigma(X_n)].\]
|
|
|
|
\end{claim*}
|
2023-06-29 20:31:46 +02:00
|
|
|
\begin{subproof}
|
2023-06-29 22:18:23 +02:00
|
|
|
\todo{TODO}
|
|
|
|
% We have $\sigma(\One_{X_{n+1} \in B}) \subseteq \sigma(X_{n}, \xi_{n+1})$.
|
2023-07-06 00:36:26 +02:00
|
|
|
% $\sigma(X_1,\ldots,X_{n-1})$
|
2023-06-29 22:18:23 +02:00
|
|
|
% is independent of $\sigma( \sigma(\One_{X_{n+1} \in B}), X_n)$.
|
2023-07-28 03:45:37 +02:00
|
|
|
% Hence the claim follows from \yaref{ceroleofindependence}.
|
2023-06-29 20:31:46 +02:00
|
|
|
\end{subproof}
|
|
|
|
\end{example}
|
|
|
|
|
2023-06-29 20:39:53 +02:00
|
|
|
{ \large\color{red}
|
2023-06-29 20:31:46 +02:00
|
|
|
New information after this point is not relevant for the exam.
|
|
|
|
}
|
|
|
|
Stopping times and optional stopping are very relevant for the exam,
|
|
|
|
the Markov property is not.
|
2023-07-04 18:09:52 +02:00
|
|
|
No notes will be allowed in the exam.
|
|
|
|
Theorems from the lecture as well as
|
|
|
|
homework exercises might be part of the exam.
|