Modelling evolutionary games as Markov chains
We can model any evolutionary game as a Markov chain by taking our state space as the set of possible populations. In doing so, we can define a transition matrix for the system. For example, for a 2-player, 2-action game we have:
where \(T_{ab} = p(a \to b)\) is the probability of transitioning from state \(a\) to state \(b\).
Types of Markov chain
Absorbing Markov chains
The first type of Markov chain is an absorbing Markov chain. This is a process where certain states can only transition to themselves, and thus the Markov chain gets stuck in said states. The transition matrix of such a process will have at least one row such that
In such a Markov chain, we can calculate an absorption matrix as follows:
- Write the transition matrix in canonical form
This is the form
Where \(Q\) is the set of transitions from transitive (non-absorbing) states to other transitive states, and \(R\) is the set of transitions from transitive states to absorbing states.
- Compute the absorption matrix
where \(B_{ab}\) is the probability of the Markov chain being absorbed into state \(b\) given that it started in state \(a\).
This can be used to analyse which strategies are favoured by the population. For example, in a Moran process, absorbing states are those in which all players have the same action type, and so the state with the highest probability of absorption would be that which has all players playing the strategy which produces the highest fitness in the model.
Ergodic Markov chains
An ergodic Markov chain is one which has no absorbing states, and so you are able to traverse from any state to any other state in a finite number of steps. In this case, we consider state distributions in the form
Where \(\pi^t_i\) is the proportion of the time spent in a certain state at step \(T\). We have that \(\sum_{i=1}^N \pi^t_i = 1\). Now, if we multiply \(\pi^t\) by the transition matrix \(T\), we get the distribution at the next step:
Thus, by taking the left eigenvector of the transition matrix \(T\), we obtain a state distribution \(\pi\) such that:
We call this the steady state of the Markov chain, and it gives us the distribution of the system amongst it's states when the system has stabilised. A Markov chain will always reach it's unique steady state provided that it is irreducible (all states can reach all other states in a finite number of steps) and aperiodic (the states do not oscillate between certain subsets). Any birth-death process fulfills these conditions, and thus reaches the steady state in a finite number of steps.