Markov chains form the basis for stochastic and dynamical systems. They describe a sequence of possible events such that the probability of each event depends only on the state attained in the previous event.

Fundamentals

Suppose we have three discrete states. Suppose also there is some probability that the state of one body either remains the same or changes to the two other states.

Some handy definitions:

Probability vector: where the entries are non-negative and sum to 1 Stochastic matrix (also *transition matrices): where each column are probability vectors; by

When we apply some stochastic matrix to some distribution , our result describes the new distribution at state . An interesting application examines our understanding of diagonalization, allowing us to explore the population as time approaches infinity. Recall that for some diagonalisable matrix, we can express it in terms of some matrix , where its columns are eigenvectors and some diagonal matrix , where its elements are eigenvalues.

We apply it to our problem, using a handy theorem. If:

Then:

This is substantially more computationally efficient — as opposed to number of computations, we instead do three. We refer to our resulting population as the equilibrium distribution for .

Example

Say we have three lecture sections and students are free to attend different lecture sections, with a probability of either staying in the lecture session they currently attend, or two probabilities of attending a different lecture section.

If we begin with some population :

And some stochastic matrix that describes their behaviour after a state change is induced:

For the first column, 70% of students that are initially in the first lecture section will stay in that section. 20% will move to the second section, and 10% will move to the third section. We observe from above, that the vector describes the behaviour at . We also observe that encodes the change between and . Our understanding of diagonalization allows us to further extend this to find the number of students in each lecture section after an infinite amount of time.

Theorems

Equilibria for regular transition matrices Let be a stochastic matrix of size .

  • There exists exactly one distribution vector such that , i.e., that is an eigenvector with eigenvalue 1. This is the equilibrium distribution for , denoted , which only has positive elements.
  • If is any distribution vector in , then .
  • is a matrix whose columns are all .