10.2. Markov processes#
Markov Process. Let \(S\) a finite set of state \(s\) that can be indexed with integer indices \(i\), \(P\) the transition probability between two states at consecutive time-steps, \(P_{ss'} = P(s'|s)\), and \(p_0\) the initial probability. A Markox process is uniquely defined by the triple
If the Markov process is time-homogeneous, state transition matrix \(P\) is independent from the time-step.
10.2.1. State transition matrix#
Given the finite set of states with integer indices, their probabilities at time \(n\) can be collected in the \(|S|\)-dimensional vector \(\mathbf{p}_n\). Transition from time \(n-1\) to \(n\) can be represented as
or using matrix formalism
As \(\mathbf{p}_n\) and \(\mathbf{P}\) represents probability, their elements are non-negative. As the \(i^{th}\) element of \(\mathbf{p}_n\) represents the probability of the system of being in state \(i\) at time \(n\), it follows
As the element \(P_{ij} = P(i|j)\) represents the conditional probability of reaching \(j\) from state \(i\) in one step, it follows
10.2.2. Properties#
10.2.2.1. Eigenvalue \(s = 1\), and stationary state \(\ \boldsymbol\pi\)#
The sum of the elements of each row of \(\mathbf{P}\) is equal to \(1\), as \(\sum_{s'} P_{ss'} = \sum_{s'} P(s' | s) = 1\). Equivalently, state transition matrix \(\mathbf{P}\) has an eigenvalue \(\lambda = 1\), with the uniform right eigenvector \(\mathbf{1}\), i.e. \(\mathbf{P} \mathbf{1} = \mathbf{1}\).
The corresponding left eigenvector is a stationary state \(\boldsymbol\pi\),
10.2.2.2. Spectrum of the transition matrix and asymptotic behavior of the system#
todo Are all the transition matrices diagonalizable?
Assuming the transition matrix can be diagonalized, the evolution of the system from the initial probability \(\mathbf{p}_0\) can be easily represented. The initial probabbility is written as \(\mathbf{p}_0 = c_k \mathbf{w}_k\), being \(\mathbf{w}_k\) the left eigenvectors of the transition matrix, and
As \(n \rightarrow +\infty\) only the contributions of the eigenvectors with \(|\lambda_k| = 1\) survive. If there’s only one of these eigenvalues, \(\lambda_1 = 1\), the system converges to \(\mathbf{1}\). If there are many eigenvalues with \(\lambda_i = 1\), there system converges with probability \(a_i\) to those stationary states, being \(a_i\) the component of the initial state of the system along the \(i^{th}\) eigenvector,
Components \(a_i\) can be evaluated as the dot product between the dual basis \(\{ \mathbf{v}_i \}_i\) w.r.t. the left eigenvectors and the initial state \(\mathbf{p}_0\), as
Existence of eigenvalues \(|s_j| = 1\), \(s_j = e^{i \theta_j}\) makes limit cycles possible as
todo Discuss the possibility (?) of negative components. What’s a negative probability?
10.2.2.3. Detailed balance#
By definition, a reversible Markov chain has a positive stationary distribution \(\boldsymbol\pi\) that satisfies the detailed balance equations
i.e. the flow between each pair of states \(i\), \(j\) is individaully balanced. Reversibility is a sufficient - even not necessary - condition for the existence of a stationary state.
Let \(\pi_i \ne 0\) for \(\forall i\), it’s possible to define a symmetric matrix via a similarity transformation, as
with \(\mathbf{S} = \mathbf{D} \mathbf{P} \mathbf{D}^{-1}\), with \(\mathbf{D} = \text{diag}\{ \sqrt{\pi_i} \}\). As the matrices \(\mathbf{P}\) and \(\mathbf{S}\) are related via a similarity transformation, they have the same eigenvalues
todo
Discuss the meaning of “reveresible”
Ergodicity. …todo start from https://en.wikipedia.org/wiki/Markov_chain#Ergodicity