In the mathematical theory of
probability
Probability is the branch of mathematics concerning numerical descriptions of how likely an Event (probability theory), event is to occur, or how likely it is that a proposition is true. The probability of an event is a number between 0 and ...
, an absorbing Markov chain is a
Markov chain
A Markov chain or Markov process is a stochastic model describing a sequence of possible events in which the probability of each event depends only on the state attained in the previous event. Informally, this may be thought of as, "What happen ...
in which every state can reach an absorbing state. An absorbing state is a state that, once entered, cannot be left.
Like general Markov chains, there can be continuous-time absorbing Markov chains with an infinite state space. However, this article concentrates on the discrete-time discrete-state-space case.
Formal definition
A Markov chain is an absorbing chain if
[
# there is at least one absorbing state and
# it is possible to go from any state to at least one absorbing state in a finite number of steps.
In an absorbing Markov chain, a state that is not absorbing is called transient.
]
Canonical form
Let an absorbing Markov chain with transition matrix ''P'' have ''t'' transient states and ''r'' absorbing states. Unlike a typical transition matrix, the rows of ''P'' represent sources, while columns represent destinations. Then
:
where ''Q'' is a ''t''-by-''t'' matrix, ''R'' is a nonzero ''t''-by-''r'' matrix, 0 is an ''r''-by-''t'' zero matrix, and ''I''''r'' is the ''r''-by-''r'' identity matrix
In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere.
Terminology and notation
The identity matrix is often denoted by I_n, or simply by I if the size is immaterial ...
. Thus, ''Q'' describes the probability of transitioning from some transient state to another while ''R'' describes the probability of transitioning from some transient state to some absorbing state.
Fundamental matrix
A basic property about an absorbing Markov chain is the expected number of visits to a transient state ''j'' starting from a transient state ''i'' (before being absorbed). The probability of transitioning from ''i'' to ''j'' in exactly ''k'' steps is the (''i'',''j'')-entry of ''Q''k. Summing this for all ''k'' (from 0 to ∞) yields the fundamental matrix, denoted by ''N''. It can be proven that
:
where ''I''''t'' is the ''t''-by-''t'' identity matrix. The (''i'', ''j'') entry of matrix ''N'' is the expected number of times the chain is in state ''j'', given
that the chain started in state ''i''. With the matrix ''N'' in hand, other properties of the Markov chain are easy to obtain.[
]
This fundamental matrix can be thought of as a matrix equivalent of the following geometric series
In mathematics, a geometric series is the sum of an infinite number of terms that have a constant ratio between successive terms. For example, the series
:\frac \,+\, \frac \,+\, \frac \,+\, \frac \,+\, \cdots
is geometric, because each su ...
:
:
Variance on number of visits
The variance on the number of visits to a transient state ''j'' with starting at a transient state ''i'' (before being absorbed) is the (''i'',''j'')-entry of the matrix
:
where ''N''dg is the diagonal matrix
In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagonal ...
with the same diagonal as ''N'' and ''N''sq is the Hadamard product of ''N'' with itself (i.e. each entry of ''N'' is squared).
Expected number of steps
The expected number of steps before being absorbed when starting in transient state ''i'' is the ''i''th entry of the vector
:
where 1 is a length-''t'' column vector whose entries are all 1.
Variance on number of steps
The variance on the number of steps before being absorbed when starting in transient state ''i'' is the ''i''th entry of the vector
:
where tsq is the Hadamard product of t with itself (i.e. each entry of t is squared).
Transient probabilities
The probability of visiting transient state ''j'' when starting at a transient state ''i'' is the (''i'',''j'')-entry of the matrix
:
where ''N''dg is the diagonal matrix
In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagonal ...
with the same diagonal as ''N''.
Absorbing probabilities
Another property is the probability of being absorbed in the absorbing state ''j'' when starting from transient state ''i'', which is the (''i'',''j'')-entry of the matrix
:
Alternatively, this probability can also be directly obtained from the (''i'',''j'')-entry of for a large enough value of ''n''. That is:
:
Examples
String generation
Consider the process of repeatedly flipping a fair coin
In probability theory and statistics, a sequence of independent Bernoulli trials with probability 1/2 of success on each trial is metaphorically called a fair coin. One for which the probability is not 1/2 is called a biased or unfair coin. In th ...
until the sequence (heads, tails, heads) appears. This process is modeled by an absorbing Markov chain with transition matrix
:
The first state represents the empty string
In formal language theory, the empty string, or empty word, is the unique string of length zero.
Formal theory
Formally, a string is a finite, ordered sequence of characters such as letters, digits or spaces. The empty string is the special cas ...
, the second state the string "H", the third state the string "HT", and the fourth state the string "HTH". Although in reality, the coin flips cease after the string "HTH" is generated, the perspective of the absorbing Markov chain is that the process has transitioned into the absorbing state representing the string "HTH" and, therefore, cannot leave.
For this absorbing Markov chain, the fundamental matrix is
:
The expected number of steps starting from each of the transient states is
:
Therefore, the expected number of coin flips before observing the sequence (heads, tails, heads) is 10, the entry for the state representing the empty string.
Games of chance
Games based entirely on chance can be modeled by an absorbing Markov chain. A classic example of this is the ancient Indian board game Snakes and Ladders. The graph on the left plots the probability mass in the lone absorbing state that represents the final square as the transition matrix is raised to larger and larger powers. To determine the expected number of turns to complete the game, compute the vector t as described above and examine tstart, which is approximately 39.2.
Infectious disease testing
Infectious disease testing, either of blood products or in medical clinics, is often taught as an example of an absorbing Markov chain. The public U.S. Centers for Disease Control and Prevention (CDC) model for HIV and for hepatitis B, for example, illustrates the property that absorbing Markov chains can lead to the detection of disease, versus the loss of detection through other means.
In the standard CDC model, the Markov chain has five states, a state in which the individual is uninfected, then a state with infected but undetectable virus, a state with detectable virus, and absorbing states of having quit/been lost from the clinic, or of having been detected (the goal). The typical rates of transition between the Markov states are the probability ''p'' per unit time of being infected with the virus, ''w'' for the rate of window period removal (time until virus is detectable), ''q'' for quit/loss rate from the system, and ''d'' for detection, assuming a typical rate at which the health system administers tests of the blood product or patients in question.
It follows that we can "walk along" the Markov model to identify the overall probability of detection for a person starting as undetected, by multiplying the probabilities of transition to each next state of the model as:
.
The subsequent total absolute number of false negative tests—the primary CDC concern—would then be the rate of tests, multiplied by the probability of reaching the infected but undetectable state, times the duration of staying in the infected undetectable state:
.
See also
* Discrete phase-type distribution
*Absorbing set (random dynamical systems)
In mathematics, an absorbing set for a random dynamical system is a subset of the phase space. A dynamical system is a system in which a function describes the time dependence of a point in a geometrical space.
The absorbing set eventually cont ...
References
{{reflist
External links
Wolfram Demonstration Project: Absorbing Markov Chain
Monopoly as a Markov chain
Markov processes
Markov models