Introduction to Markov chain mixing
Recall
if ($X_n)_n$ is an irreducible Markov chain with stationary distribution $\pi$, thhen
$\lim_{n \to \infty} \frac{1}{n} \sum_{j=0}^n 1_{[X_j = x]} = \pi(x)$, $ \quad \mathbb{P}_\mu$ - a.s
Today's goal
We will show that $X_n$ converges to $\pi$ under some "strong sense".
- Total variation distance
- The convergence theorem
- Mixing times
There is three ways to characterize the total variation distance:
$\mu$ and $\nu$ : probability measures on $\Omega$.
$\mid\mid \mu - \nu \mid\mid_{TV}$ = $max_{A\subset \Omega} \mid\mu (A) - \nu (A)\mid$.
Lemma
- $\mid\mid \mu - \nu \mid\mid_{TV} = \frac{1}{2} \sum_{x \in \Omega} \mid \mu (x) - \nu (x)\mid$.
- $\mid\mid \mu - \nu \mid\mid_{TV} = \frac{1}{2} sup{\mu \f - \nu \f : \f satisfying \max_{x\in \Omega} \mid \f(x) \mid \leq 1}$.
- $\mid\mid \mu - \nu \mid\mid_{TV} = inf{\quad \mathbb{P} [X \neq Y] : (X, Y) is a coupling of \mu , \nu }.
Definition
We call (X, Y) the optimal coupling if $\quad \mathbb{P} [X \neq Y] = \mid\mid \mu - \nu \mid\mid_{TV}$.
The Convergence Theorem
Suppose that ($X_n)_n$ is a Markov chain with transition matrix P. Assume that P is irreducible and aperiodic, then
- there exists r such that $\mathbb{P}^r (x, y) > 0 for all x, y \in \Omega$;
- there exists a unique stationary distribution $\pi and \pi(x) > 0 for all x \in \Omega$.
Theorem
Suppose that P is irreducible, aperiodic, with stationary distribution $\pi$.
Then there exist constants $\alpha \in$ (0, 1) and C > 0 such that
$\max_{x\in \Omega}\mid\mid \mathbb{P}^n (x, .) - \pi \mid\mid_{TV} \leq \mathbb{C}\alpha ^n \ \forall n \geq 1$.
Mixing time
Definition
d(n) = $\max_{x\in \Omega}\mid\mid \mathbb{p}^n (x, .) - \pi \mid\mid_{TV}$
$\bar d (n) = \max_{x,y \in \Omega} \mid\mid \mathbb{P}^n(x,.) - \mathbb{P}^n (y,.) \mid\mid_{TV}$
Lemma
$ d(n) \leq \bar d (n) \leq 2d(n)$
Lemma
$\bar d (m + n) \leq \bar d (m) . \bar d (n)$
Corollary
$\bar d (mn) \leq \bar d (n)^m$
Definition
$ t_{mix} = \min {n : d(n) \leq \frac{1}{4} }, \ \ t_{mix} (\in) = \min {n : d(n) \leq \in}$
Lemma
$ t_{mix} (\in) \leq$ $\log({\frac{1}{\in}}) \frac{t_{mix}}{\log{2}}$
Couple tow Markov chains
Definition
A coupling of two Markov chains with transition matrix P is a process ($\X_n, \Y_n)_{n \geq 0}$ with the following two properties.
- Both ($\X_n ) and (\Y_n)$ are Markov chains with transition matrix P.
- They stay together after their first meet.
Theorem
Suppose that P is irreducible with stationary distribution $\pi$. Let $(\X_n, \Y_n)_{n \geq 0}$ be a coupling of Markov chains with transition matrix P for which $\X_0 = x, \Y_0 = y$. Define $\tau$ to be their first meet time :
$\tau = \min {n \geq 0 : \X_n = \Y_n }$.
Then
$\mid\mid \mathbb{P}^n (x, .) - \mathbb{P}^n (y, .)\mid\mid_{TV} \leq \quad\mathbb{P}_{x,y} [\tau \> n]$.
In particular,
d(n) $\leq \max_{x,y} \quad\mathbb{P}_{x,y}[\tau > n]$.
Random walk on N-cycle : Upper bound on $t_{mix}$
Lazy walk : it remains in current position with probability 1 / 2, moves left with probability 1 / 4, right with probability 1 / 4.
- It is irreducible.
- The stationary measure is the uniform measure.
Theorem
For the lazy walk on N-cycle, we have
$ t_{mix} \leq N^2$.
Comments
Post a Comment