Discrete-time Markov chain
   HOME

TheInfoList



OR:

In
probability Probability is the branch of mathematics concerning numerical descriptions of how likely an event is to occur, or how likely it is that a proposition is true. The probability of an event is a number between 0 and 1, where, roughly speakin ...
, a discrete-time Markov chain (DTMC) is a sequence of random variables, known as a stochastic process, in which the value of the next variable depends only on the value of the current variable, and not any variables in the past. For instance, a machine may have two states, ''A'' and ''E''. When it is in state ''A'', there is a 40% chance of it moving to state ''E'' and a 60% chance of it remaining in state ''A''. When it is in state ''E'', there is a 70% chance of it moving to ''A'' and a 30% chance of it staying in ''E''. The sequence of states of the machine is a Markov chain. If we denote the chain by X_0, X_1, X_2, ... then X_0 is the state which the machine starts in and X_ is the random variable describing its state after 10 transitions. The process continues forever, indexed by the natural numbers. An example of a stochastic process which is not a Markov chain is the model of a machine which has states ''A'' and ''E'' and moves to ''A'' from either state with 50% chance if it has ever visited ''A'' before, and 20% chance if it has never visited ''A'' before (leaving a 50% or 80% chance that the machine moves to ''E''). This is because the behavior of the machine depends on the whole history—if the machine is in ''E'', it may have a 50% or 20% chance of moving to ''A'', depending on its past values. Hence, it does not have the
Markov property In probability theory and statistics, the term Markov property refers to the memoryless property of a stochastic process. It is named after the Russian mathematician Andrey Markov. The term strong Markov property is similar to the Markov propert ...
. A Markov chain can be described by a
stochastic matrix In mathematics, a stochastic matrix is a square matrix used to describe the transitions of a Markov chain. Each of its entries is a nonnegative real number representing a probability. It is also called a probability matrix, transition matrix, ...
, which lists the probabilities of moving to each state from any individual state. From this matrix, the probability of being in a particular state ''n'' steps in the future can be calculated. A Markov chain's state space can be partitioned into communicating classes that describe which states are reachable from each other (in one transition or in many). Each state can be described as transient or recurrent, depending on the probability of the chain ever returning to that state. Markov chains can have properties including periodicity, reversibility and stationarity. A
continuous-time Markov chain A continuous-time Markov chain (CTMC) is a continuous stochastic process in which, for each state, the process will change state according to an exponential random variable and then move to a different state as specified by the probabilities of ...
is like a discrete-time Markov chain, but it moves states continuously through time rather than as discrete time steps. Other stochastic processes can satisfy the Markov property, the property that past behavior does not affect the process, only the present state.


Definition

A discrete-time Markov chain is a sequence of random variables X_0, X_1, X_2, ... with the
Markov property In probability theory and statistics, the term Markov property refers to the memoryless property of a stochastic process. It is named after the Russian mathematician Andrey Markov. The term strong Markov property is similar to the Markov propert ...
, namely that the probability of moving to the next state depends only on the present state and not on the previous states: :\Pr(X_=x\mid X_1=x_1, X_2=x_2, \ldots, X_n=x_n) = \Pr(X_=x\mid X_n=x_n), if both
conditional probabilities In probability theory, conditional probability is a measure of the probability of an event occurring, given that another event (by assumption, presumption, assertion or evidence) has already occurred. This particular method relies on event B occur ...
are well defined, that is, if \Pr(X_1=x_1,\ldots,X_n=x_n)>0. The possible values of ''X''''i'' form a countable set ''S'' called the state space of the chain. Markov chains are often described by a sequence of
directed graph In mathematics, and more specifically in graph theory, a directed graph (or digraph) is a graph that is made up of a set of vertices connected by directed edges, often called arcs. Definition In formal terms, a directed graph is an ordered pa ...
s, where the edges of graph ''n'' are labeled by the probabilities of going from one state at time ''n'' to the other states at time ''n'' + 1, \Pr(X_=x\mid X_n=x_n). The same information is represented by the transition matrix from time ''n'' to time ''n'' + 1. However, Markov chains are frequently assumed to be time-homogeneous (see variations below), in which case the graph and matrix are independent of ''n'' and are thus not presented as sequences. These descriptions highlight the structure of the Markov chain that is independent of the initial distribution \Pr(X_1=x_1). When time-homogeneous, the chain can be interpreted as a state machine assigning a probability of hopping from each vertex or state to an adjacent one. The probability \Pr(X_n=x\mid X_1=x_1) of the machine's state can be analyzed as the statistical behavior of the machine with an element x_1 of the state space as input, or as the behavior of the machine with the initial distribution \Pr(X_1=y)= _1=y/math> of states as input, where /math> is the Iverson bracket.


Variations

*Time-homogeneous Markov chains (or stationary Markov chains) are processes where ::\Pr(X_=x\mid X_n=y) = \Pr(X_n=x\mid X_=y) :for all ''n''. The probability of the transition is independent of ''n''. *A Markov chain with memory (or a Markov chain of order ''m'') :where ''m'' is finite, is a process satisfying :: \begin &\Pr(X_n=x_n\mid X_=x_, X_=x_, \dots , X_1=x_1) \\ = &\Pr(X_n=x_n\mid X_=x_, X_=x_, \dots, X_=x_) \textn > m \end :In other words, the future state depends on the past ''m'' states. It is possible to construct a chain (Y_n) from (X_n) which has the 'classical' Markov property by taking as state space the ordered ''m''-tuples of ''X'' values, ie. Y_n= \left( X_n,X_,\ldots,X_ \right).


''n''-step transitions

The probability of going from state ''i'' to state ''j'' in ''n'' time steps is :p_^ = \Pr(X_n=j\mid X_0=i) and the single-step transition is :p_ = \Pr(X_1=j\mid X_0=i). For a time-homogeneous Markov chain: :p_^ = \Pr(X_=j \mid X_=i) and :p_ = \Pr(X_=j \mid X_k=i). The ''n''-step transition probabilities satisfy the Chapman–Kolmogorov equation, that for any ''k'' such that 0 < ''k'' < ''n'', :p_^ = \sum_ p_^ p_^ where ''S'' is the state space of the Markov chain. The
marginal distribution In probability theory and statistics, the marginal distribution of a subset of a collection of random variables is the probability distribution of the variables contained in the subset. It gives the probabilities of various values of the varia ...
Pr(''X''''n'' = ''x'') is the distribution over states at time ''n''. The initial distribution is Pr(''X''0 = ''x''). The evolution of the process through one time step is described by : \Pr(X_n=j) = \sum_ p_ \Pr(X_=r) = \sum_ p_^ \Pr(X_0=r). (The superscript (''n'') is an index, and not an exponent).


Communicating classes and properties

A state ''j'' is said to be accessible from a state ''i'' (written ''i'' → ''j'') if a system started in state ''i'' has a non-zero probability of transitioning into state ''j'' at some point. Formally, state ''j'' is accessible from state ''i'' if there exists an integer ''nij'' ≥ 0 such that : \Pr(X_=j \mid X_0=i) = p_^ > 0. A state ''i'' is said to communicate with state ''j'' (written ''i'' ↔ ''j'') if both ''i'' → ''j'' and ''j'' → ''i''. A communicating class is a maximal set of states ''C'' such that every pair of states in ''C'' communicates with each other. Communication is an equivalence relation, and communicating classes are the equivalence classes of this relation. A communicating class is closed if the probability of leaving the class is zero, namely if ''i'' is in ''C'' but ''j'' is not, then ''j'' is not accessible from ''i''. The set of communicating classes forms a directed, acyclic graph by inheriting the arrows from the original state space. A communicating class is closed if and only if it has no outgoing arrows in this graph. A state ''i'' is said to be essential or final if for all ''j'' such that ''i'' → ''j'' it is also true that ''j'' → ''i''. A state ''i'' is inessential if it is not essential. A state is final if and only if its communicating class is closed. A Markov chain is said to be irreducible if its state space is a single communicating class; in other words, if it is possible to get to any state from any state.


Periodicity

A state i has period k if any return to state i must occur in multiples of k time steps. Formally, the
period Period may refer to: Common uses * Era, a length or span of time * Full stop (or period), a punctuation mark Arts, entertainment, and media * Period (music), a concept in musical composition * Periodic sentence (or rhetorical period), a concept ...
of state i is defined as : k = \gcd\ (where \gcd is the
greatest common divisor In mathematics, the greatest common divisor (GCD) of two or more integers, which are not all zero, is the largest positive integer that divides each of the integers. For two integers ''x'', ''y'', the greatest common divisor of ''x'' and ''y'' is ...
) provided that this set is not empty. Otherwise the period is not defined. Note that even though a state has period k, it may not be possible to reach the state in k steps. For example, suppose it is possible to return to the state in \ time steps; k would be 2, even though 2 does not appear in this list. If k = 1, then the state is said to be aperiodic. Otherwise ( k > 1), the state is said to be periodic with period k. Periodicity is a class property—that is, if a state has period k then every state in its communicating class has period k.


Transience and recurrence

A state ''i'' is said to be transient if, given that we start in state ''i'', there is a non-zero probability that we will never return to ''i''. Formally, let the random variable ''Ti'' be the first return time to state ''i'' (the "hitting time"): : T_i = \inf \. The number : f_^ = \Pr(T_i = n \mid X_0 = i) is the probability that we return to state ''i'' for the first time after ''n'' steps. Therefore, state ''i'' is transient if : \Pr(T_i < \infty \mid X_0 = i) = \sum_^\infty f_^ < 1. State ''i'' is recurrent (or persistent) if it is not transient. Recurrence and transience are class properties, that is, they either hold or do not hold equally for all members of a communicating class. A state ''i'' is recurrent
if and only if In logic and related fields such as mathematics and philosophy, "if and only if" (shortened as "iff") is a biconditional logical connective between statements, where either both statements are true or both are false. The connective is b ...
the expected number of visits to ''i'' is infinite: :\sum_^\infty p_^ = \infty.


Positive recurrence

Even if the hitting time is finite with probability ''1'', it need not have a finite expectation. The mean recurrence time at state ''i'' is the expected return time ''Mi'': : M_i = E _i\sum_^\infty n\cdot f_^. State ''i'' is positive recurrent (or non-null persistent) if ''Mi'' is finite; otherwise, state ''i'' is null recurrent (or null persistent). Positive and null recurrence are classes properties.


Absorbing states

A state ''i'' is called absorbing if it is impossible to leave this state. Therefore, the state ''i'' is absorbing if and only if : p_ = 1\textp_ = 0\texti \not= j. If every state can reach an absorbing state, then the Markov chain is an
absorbing Markov chain In the mathematical theory of probability, an absorbing Markov chain is a Markov chain in which every state can reach an absorbing state. An absorbing state is a state that, once entered, cannot be left. Like general Markov chains, there can be c ...
.


Reversible Markov chain

A Markov chain is said to be reversible if there is a probability distribution over its states such that :\pi_i \Pr(X_ = j \mid X_ = i) = \pi_j \Pr(X_ = i \mid X_ = j) for all times ''n'' and all states ''i'' and ''j''. This condition is known as the
detailed balance The principle of detailed balance can be used in kinetic systems which are decomposed into elementary processes (collisions, or steps, or elementary reactions). It states that at equilibrium, each elementary process is in equilibrium with its reve ...
condition (or local balance equation). Considering a fixed arbitrary time ''n'' and using the shorthand :p_ = \Pr(X_ = j \mid X_n = i)\,, the detailed balance equation can be written more compactly as :\pi_i p_ = \pi_j p_\,. The single time-step from ''n'' to ''n'' + 1 can be thought of as each person ''i'' having ''i'' dollars initially and paying each person ''j'' a fraction ''pij'' of it. The detailed balance condition states that upon each payment, the other person pays exactly the same amount of money back. Clearly the total amount of money each person has remains the same after the time-step, since every dollar spent is balanced by a corresponding dollar received. This can be shown more formally by the equality :\sum_i \pi_i p_ = \sum_i \pi_j p_ = \pi_j \sum_i p_ = \pi_j\,, which essentially states that the total amount of money person ''j'' receives (including from himself) during the time-step equals the amount of money he pays others, which equals all the money he initially had because it was assumed that all money is spent (that is, ''pji'' sums to 1 over ''i''). The assumption is a technical one, because the money not really used is simply thought of as being paid from person ''j'' to himself (that is, ''pjj'' is not necessarily zero). As ''n'' was arbitrary, this reasoning holds for any ''n'', and therefore for reversible Markov chains is always a steady-state distribution of Pr(''X''''n''+1 = ''j'' ,  ''X''''n'' = ''i'') for every ''n''. If the Markov chain begins in the steady-state distribution, that is, if \Pr(X_0=i)=\pi_i, then \Pr(X_n=i)=\pi_i for all n and the detailed balance equation can be written as :\Pr(X_ = i, X_ = j) = \Pr(X_ = i, X_ = j)\,. The left- and right-hand sides of this last equation are identical except for a reversing of the time indices ''n'' and ''n'' + 1.
Kolmogorov's criterion In probability theory, Kolmogorov's criterion, named after Andrey Kolmogorov, is a theorem giving a necessary and sufficient condition for a Markov chain or continuous-time Markov chain to be stochastically identical to its time-reversed version. ...
gives a necessary and sufficient condition for a Markov chain to be reversible directly from the transition matrix probabilities. The criterion requires that the products of probabilities around every closed loop are the same in both directions around the loop. Reversible Markov chains are common in Markov chain Monte Carlo (MCMC) approaches because the detailed balance equation for a desired distribution necessarily implies that the Markov chain has been constructed so that is a steady-state distribution. Even with time-inhomogeneous Markov chains, where multiple transition matrices are used, if each such transition matrix exhibits detailed balance with the desired distribution, this necessarily implies that is a steady-state distribution of the Markov chain.


Closest reversible Markov chain

For any time-homogeneous Markov chain given by a transition matrix P \in \mathbb^, any norm , , \cdot , , on \mathbb^ which is induced by a
scalar product In mathematics, the dot product or scalar productThe term ''scalar product'' means literally "product with a scalar as a result". It is also used sometimes for other symmetric bilinear forms, for example in a pseudo-Euclidean space. is an alge ...
, and any probability vector \pi, there exists a unique transition matrix P^* which is reversible according to \pi and which is closest to P according to the norm , , \cdot , , . The matrix P^* can be computed by solving a quadratic-convex optimization problem. For example, consider the following Markov chain: This Markov chain is not reversible. According to the Frobenius Norm the closest reversible Markov chain according to \pi = \left(\frac,\frac,\frac\right) can be computed as If we choose the probability vector randomly as \pi=\left( \frac, \frac, \frac \right), then the closest reversible Markov chain according to the Frobenius norm is approximately given by


Stationary distributions

A distribution \pi is a stationary distribution of the Markov chain with stochastic matrix P if and only if \pi P = \pi. This can be written as: :\forall j\in \mathbb: \sum_ \pi_i p_ = \pi_j This condition implies that \pi P^n=\pi and hence if the Markov chain (X_n, n\in \mathbb) has initial distribution X_0 = \pi then X_n = \pi (in distribution) for any n\in\mathbb. If a Markov chain is irreducible then it has a stationary distribution if and only if it is positive recurrent, in which case the unique such distribution is given by \pi_i=\frac where M_i=\mathbb(T_i) is the mean recurrence time of ''i''. If a chain has more than one closed communicating class, its stationary distributions will not be unique (consider any closed communicating class C_i in the chain; each one will have its own unique stationary distribution \pi_i. Extending these distributions to the overall chain, setting all values to zero outside the communication class, yields that the set of invariant measures of the original chain is the set of all convex combinations of the \pi_i's). However, if a state ''j'' is aperiodic, then \lim\nolimits_ p_^ = \tfrac and for any other state ''i'', let ''fij'' be the probability that the chain ever visits state ''j'' if it starts at ''i'', \lim\nolimits_ p_^ = C \tfrac. If a state ''i'' is periodic with period ''k'' > 1 then the limit \lim\nolimits_ p_^ does not exist, although the limit \lim\nolimits_ p_^ does exist for every integer ''r''.


Steady-state analysis and the time-inhomogeneous Markov chain

A Markov chain need not necessarily be time-homogeneous to have an equilibrium distribution. If there is a probability distribution over states \boldsymbol such that :\pi_j = \sum_ \pi_i \, \Pr(X_ = j \mid X_n = i) for every state ''j'' and every time ''n'' then \boldsymbol is an equilibrium distribution of the Markov chain. Such can occur in Markov chain Monte Carlo (MCMC) methods in situations where a number of different transition matrices are used, because each is efficient for a particular kind of mixing, but each matrix respects a shared equilibrium distribution.


Hitting times

The hitting time is the time, starting in a given set of states until the chain arrives in a given state or set of states. The distribution of such a time period has a phase type distribution. The simplest such distribution is that of a single exponentially distributed transition.


Expected hitting times

For a subset of states ''A'' ⊆ ''S'', the vector ''k''''A'' of hitting times (where element k_i^A represents the expected value, starting in state ''i'' that the chain enters one of the states in the set ''A'') is the minimal non-negative solution to :\begin k_i^A = 0 & \text i \in A\\ -\sum_ q_ k_j^A = 1&\text i \notin A. \end


Ergodic theorem

An instance of ergodic theory, the ergodic theorem for states that for an irreducible aperiodic Markov chain, with any two states ''i'' and ''j'', :p_^\rightarrow \frac as n\rightarrow \infty


Notes


References

* A. A. Markov (1971). "Extension of the limit theorems of probability theory to a sum of variables connected in a chain". reprinted in Appendix B of: R. Howard. ''Dynamic Probabilistic Systems, volume 1: Markov Chains''. John Wiley and Sons. * * Leo Breiman (1992)
968 Year 968 ( CMLXVIII) was a leap year starting on Wednesday (link will display the full calendar) of the Julian calendar. Events By place Byzantine Empire * Emperor Nikephoros II receives a Bulgarian embassy led by Prince Boris (th ...
''Probability''. Original edition published by Addison-Wesley; reprinted by Society for Industrial and Applied Mathematics . (See Chapter 7) *
J. L. Doob Joseph Leo Doob (February 27, 1910 – June 7, 2004) was an United States of America, American mathematician, specializing in Mathematical analysis, analysis and probability theory. The theory of Martingale (probability theory), martingales was ...
(1953) ''Stochastic Processes''. New York: John Wiley and Sons . * S. P. Meyn and R. L. Tweedie (1993) ''Markov Chains and Stochastic Stability''. London: Springer-Verlag . online
MCSS
. Second edition to appear, Cambridge University Press, 2009. * Classical text. cf Chapter 6 ''Finite Markov Chains'' pp. 384ff. * John G. Kemeny & J. Laurie Snell (1960) ''Finite Markov Chains'', D. van Nostrand Company * E. Nummelin. "General irreducible Markov chains and non-negative operators". Cambridge University Press, 1984, 2004. * Seneta, E. ''Non-negative matrices and Markov chains''. 2nd rev. ed., 1981, XVI, 288 p., Softcover Springer Series in Statistics. (Originally published by Allen & Unwin Ltd., London, 1973) {{refend zh-yue:離散時間馬可夫鏈 Markov processes