Markov models—Markov chains

You can look back there to explain things, but the explanation disappears. You’ll never find it there. Things are not explained by the past. They’re explained by what happens now. –Alan Watts

Access optionsAccess options

Rent or Buy article

Get time limited or full article access on ReadCube.


All prices are NET prices.

Fig. 1: State transition models, transition matrices T, and the number of transitions required to approximate the steady-state limiting distributions, Tn (n→∞), to the displayed number of decimal places.
Fig. 2: Effect of the initial state (G, M) on the state evolution of 5,000 Markov chains with 20% and 40% chances of arrest.


  1. 1.

    Skewes, A. D. & Welch, R. D. PeerJ 1, e127 (2013).

Download references

Author information

Correspondence to Martin Krzywinski.

Ethics declarations

Competing interests

The authors declare no competing interests.

Rights and permissions

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark