Segregating Markov Chains
Artikel i vetenskaplig tidskrift, 2018
Dealing with finite Markov chains in discrete time, the focus often lies on convergence behavior and one tries to make different copies of the chain meet as fast as possible and then stick together. There are, however, discrete finite (reducible) Markov chains, for which two copies started in different states can be coupled to meet almost surely in finite time, yet their distributions keep a total variation distance bounded away from 0, even in the limit as time tends to infinity. We show that the supremum of total variation distance kept in this context is .
Total variation distance