Continuous-time Markov chain
   HOME

TheInfoList



OR:

A continuous-time Markov chain (CTMC) is a continuous
stochastic process In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables. Stochastic processes are widely used as mathematical models of systems and phenomena that ap ...
in which, for each state, the process will change state according to an exponential random variable and then move to a different state as specified by the probabilities of a stochastic matrix. An equivalent formulation describes the process as changing state according to the least value of a set of exponential random variables, one for each possible state it can move to, with the parameters determined by the current state. An example of a CTMC with three states \ is as follows: the process makes a transition after the amount of time specified by the holding time—an exponential random variable E_i, where ''i'' is its current state. Each random variable is independent and such that E_0\sim \text(6), E_1\sim \text(12) and E_2\sim \text(18). When a transition is to be made, the process moves according to the jump chain, a
discrete-time Markov chain In probability, a discrete-time Markov chain (DTMC) is a sequence of random variables, known as a stochastic process, in which the value of the next variable depends only on the value of the current variable, and not any variables in the past. Fo ...
with stochastic matrix: :\begin 0 & \frac & \frac \\ \frac & 0 & \frac \\ \frac & \frac & 0 \end. Equivalently, by the property of competing exponentials, this CTMC changes state from state ''i'' according to the minimum of two random variables, which are independent and such that E_\sim \text(q_) for i\neq j where the parameters are given by the Q-matrix Q=(q_) :\begin -6 & 3 & 3 \\ 4 & -12 & 8 \\ 15 & 3 & -18 \end. Each non-diagonal entry q_ can be computed as the probability that the jump chain moves from state ''i'' to state ''j'', divided by the expected holding time of state ''i''. The diagonal entries are chosen so that each row sums to 0. A CTMC satisfies the Markov property, that its behavior depends only on its current state and not on its past behavior, due to the memorylessness of the exponential distribution and of discrete-time Markov chains.


Definition

Let (\Omega,,\Pr) be a probability space, let S be a countable nonempty set, and let T=\mathbb R_ (T for "time"). Equip S with the
discrete metric Discrete may refer to: *Discrete particle or quantum in physics, for example in quantum theory *Discrete device, an electronic component with just one circuit element, either passive or active, other than an integrated circuit *Discrete group, a g ...
, so that we can make sense of right continuity of functions \mathbb R_\to S. A continuous-time Markov chain is defined by: * A probability vector \lambda on S (which below we will interpret as the ''initial distribution'' of the Markov chain), and * A rate matrix Q on S, that is, a function Q:S^2\to\mathbb R such that # for all distinct i,j\in S, 0\le Q_, # for all i\in S, \sum_Q_=-Q_. (Even if I is infinite, this sum is ''a priori'' well defined (possibly equalling +\infty) because each term appearing in the sum is nonnegative. ''A posteriori'', we know the sum must also be finite (not equal to +\infty), since we're assuming it's equal to -Q_ and we've assumed Q is real valued. Some authors instead use a definition that's word-for-word the same except for a modified stipulation Q:S^2\to\mathbb R\cup\, and say Q is ''stable'' or ''totally stable'' to mean \operatornameQ\subseteq\mathbb R, i.e., every entry is real valued.) Note that the row sums of Q are 0: \forall i\in S~\sum_Q_=0, or more succinctly, Q\cdot1=0. This situation contrasts with the situation for
discrete-time Markov chain In probability, a discrete-time Markov chain (DTMC) is a sequence of random variables, known as a stochastic process, in which the value of the next variable depends only on the value of the current variable, and not any variables in the past. Fo ...
s, where all row sums of the transition matrix equal unity. Now, let X:T\to S^\Omega such that \forall t\in T~X(t) is (,S)-measurable. There are three equivalent ways to define X being ''Markov with initial distribution \lambda and rate matrix Q'': via transition probabilities or via the jump chain and holding times. As a prelude to a transition-probability definition, we first motivate the definition of a ''regular'' rate matrix. We will use the transition rate matrix Q to specify the dynamics of the Markov chain by means of generating a collection of ''transition matrices'' P(t) on S (t\in\mathbb R_), via the following theorem. We say Q is ''regular'' to mean that we do have uniqueness for the above system, i.e., that there exists exactly one solution. We say Q is ''irregular'' to mean Q is not regular. If S is finite, then there is exactly one solution, namely P=(e^)_, and hence Q is regular. Otherwise, S is infinite, and there exist irregular transition rate matrices on S. If Q is regular, then for the unique solution P, for each t\in T, P(t) will be a stochastic matrix. We will assume Q is regular from the beginning of the following subsection up through the end of this section, even though it is conventional to not include this assumption. (Note for the expert: thus we are not defining continuous-time Markov chains in general but only ''non-explosive'' continuous-time Markov chains.)


Transition-probability definition

Let P be the (unique) solution of the system (). (Uniqueness guaranteed by our assumption that Q is regular.) We say X is ''Markov with initial distribution \lambda and rate matrix Q'' to mean: for any nonnegative integer n\ge0, for all t_0,\dots,t_\in T such that t_0<\dots for all i_0,\dots,i_\in I, Using induction and the fact that \forall A,B\in~~\Pr(B)\ne0\rightarrow\Pr(A\cap B)=\Pr(A\mid B)\Pr(B), we can show the equivalence of the above statement containing () and the following statement: for all i\in I,~\Pr(X_0=i)=\lambda_i and for any nonnegative integer n\ge0, for all t_0,\dots,t_\in T such that t_0<\dots for all i_0,\dots,i_\in I such that 0<\Pr(X_0=i_0,\dots,X_=i_n) (it follows that 0<\Pr(X_=i_n)), It follows from continuity of the functions (P(t)_)_ (i,j\in S) that the trajectory (X_t(\omega))_ is almost surely right continuous (with respect to the
discrete metric Discrete may refer to: *Discrete particle or quantum in physics, for example in quantum theory *Discrete device, an electronic component with just one circuit element, either passive or active, other than an integrated circuit *Discrete group, a g ...
on S): there exists a \Pr-null set N such that \\subseteq N.


Jump-chain/holding-time definition


Sequences associated to a right-continuous function

Let f:T\to S be right continuous (when we equip S with the
discrete metric Discrete may refer to: *Discrete particle or quantum in physics, for example in quantum theory *Discrete device, an electronic component with just one circuit element, either passive or active, other than an integrated circuit *Discrete group, a g ...
). Define :h=h(f)=(\inf\)_)\cup\, let :H=H(f)=(h^0)_ be the ''holding-time sequence'' associated to f, choose s\in S, and let :y=y(f)=\left(\beginf(\sum_H_k)&\text\sum_H_k<+\infty,\\s&\text\end\right)_ be "the ''state sequence''" associated to f.


Definition of the jump matrix Π

The jump matrix \Pi, alternatively written \Pi(Q) if we wish to emphasize the dependence on Q, is the matrix \Pi=( =j_\cup\bigcup_(\\cup\), where Z=Z(Q)=\ is the ''zero set'' of the function (Q_)_.


Jump-chain/holding-time property

We say X is ''Markov with initial distribution \lambda and rate matrix Q'' to mean: the trajectories of X are almost surely right continuous, let f be a modification of X to have (everywhere) right-continuous trajectories, \sum_H(f(\omega))_n=+\infty almost surely (note to experts: this condition says X is non-explosive), the state sequence y(f(\omega)) is a discrete-time Markov chain with initial distribution \lambda (jump-chain property) and transition matrix \Pi(Q), and \forall n\in\mathbb Z_~\forall B\in(\mathbb R_)~\Pr(H_n(f)\in B)=\operatorname(-Q_)(B) (holding-time property).


Infinitesimal definition

We say X is ''Markov with initial distribution \lambda and rate matrix Q'' to mean: for all i\in S, \Pr(X(0)=i)=\lambda_i and for all i,j, for all t and for small strictly positive values of h, the following holds for all t\in T such that 0<\Pr(X(t)=i): :\Pr(X(t+h) = j \mid X(t) = i) = =j+ Q_h + o(h), where the little-o term o(h) depends in a certain way on i,j,h. The above equation shows that q_ can be seen as measuring how quickly the transition from i to j happens for i\neq j, and how quickly the transition away from i happens for i= j.


Properties


Communicating classes

Communicating classes, transience, recurrence and positive and null recurrence are defined identically as for
discrete-time Markov chain In probability, a discrete-time Markov chain (DTMC) is a sequence of random variables, known as a stochastic process, in which the value of the next variable depends only on the value of the current variable, and not any variables in the past. Fo ...
s.


Transient behaviour

Write P(''t'') for the matrix with entries ''p''''ij'' = P(''X''''t'' = ''j'' ,  ''X''0 = ''i''). Then the matrix P(''t'') satisfies the forward equation, a
first-order differential equation In mathematics, an ordinary differential equation (ODE) is a differential equation whose unknown(s) consists of one (or more) function(s) of one variable and involves the derivatives of those functions. The term ''ordinary'' is used in contrast ...
:P'(t) = P(t) Q where the prime denotes differentiation with respect to ''t''. The solution to this equation is given by a
matrix exponential In mathematics, the matrix exponential is a matrix function on square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exponential give ...
:P(t) = e^ In a simple case such as a CTMC on the state space . The general ''Q'' matrix for such a process is the following 2 × 2 matrix with ''α'',''β'' > 0 :Q = \begin -\alpha & \alpha \\ \beta & -\beta \end. The above relation for forward matrix can be solved explicitly in this case to give :P(t) = \begin \frac + \frace^ & \frac - \frace^ \\ \frac - \frace^ & \frac + \frace^ \end However, direct solutions are complicated to compute for larger matrices. The fact that ''Q'' is the generator for a
semigroup In mathematics, a semigroup is an algebraic structure consisting of a set together with an associative internal binary operation on it. The binary operation of a semigroup is most often denoted multiplicatively: ''x''·''y'', or simply ''xy'', ...
of matrices :P(t+s) = e^ = e^ e^ = P(t) P(s) is used.


Stationary distribution

The stationary distribution for an irreducible recurrent CTMC is the probability distribution to which the process converges for large values of ''t''. Observe that for the two-state process considered earlier with P(''t'') given by :P(t) = \begin \frac + \frace^ & \frac - \frace^ \\ \frac - \frace^ & \frac + \frace^ \end as ''t'' → ∞ the distribution tends to :P_\pi = \begin \frac & \frac \\ \frac & \frac \end Observe that each row has the same distribution as this does not depend on starting state. The row vector ' may be found by solving :\pi Q = 0. with the additional constraint that :\sum_ \pi_i = 1.


Example 1

The image to the right describes a continuous-time Markov chain with state-space and
transition rate matrix Transition or transitional may refer to: Mathematics, science, and technology Biology * Transition (genetics), a point mutation that changes a purine nucleotide to another purine (A ↔ G) or a pyrimidine nucleotide to another pyrimidine (C ↔ ...
:: Q=\begin -0.025 & 0.02 & 0.005 \\ 0.3 & -0.5 & 0.2 \\ 0.02 & 0.4 & -0.42 \end. The stationary distribution of this chain can be found by solving \pi Q=0, subject to the constraint that elements must sum to 1 to obtain :\pi = \begin0.885 & 0.071 & 0.044 \end.


Example 2

The image to the right describes a discrete-time Markov chain modeling
Pac-Man originally called ''Puck Man'' in Japan, is a 1980 maze action video game developed and released by Namco for arcades. In North America, the game was released by Midway Manufacturing as part of its licensing agreement with Namco America. Th ...
with state-space . The player controls Pac-Man through a maze, eating pac-dots. Meanwhile, he is being hunted by ghosts. For convenience, the maze shall be a small 3x3-grid and the monsters move randomly in horizontal and vertical directions. A secret passageway between states 2 and 8 can be used in both directions. Entries with probability zero are removed in the following transition rate matrix: Q=\begin -1 & \frac & & \frac\\ \frac & -1 & \frac & & \frac&&&\frac\\ & \frac & -1 & & & \frac\\ \frac & & & -1 & \frac & & \frac\\ & \frac & & \frac & -1 & \frac & & \frac\\ & & \frac & & \frac & -1& & & \frac\\ & & & \frac & & & -1 & \frac\\ & \frac & && \frac & & \frac & -1 & \frac\\ & & & & & \frac & & \frac & -1\end This Markov chain is irreducible, because the ghosts can fly from every state to every state in a finite amount of time. Due to the secret passageway, the Markov chain is also aperiodic, because the monsters can move from any state to any state both in an even and in an uneven number of state transitions. Therefore, a unique stationary distribution exists and can be found by solving \pi Q=0, subject to the constraint that elements must sum to 1. The solution of this linear equation subject to the constraint is \pi=(7.7,15.4,7.7,11.5,15.4,11.5,7.7,15.4,7.7)\%. The central state and the border states 2 and 8 of the adjacent secret passageway are visited most and the corner states are visited least.


Time reversal

For a CTMC ''X''''t'', the time-reversed process is defined to be \hat X_t = X_. By
Kelly's lemma In probability theory, Kelly's lemma states that for a stationary continuous-time Markov chain, a process defined as the time-reversed process has the same stationary distribution as the forward-time process. The theorem is named after Frank Kelly ...
this process has the same stationary distribution as the forward process. A chain is said to be reversible if the reversed process is the same as the forward process.
Kolmogorov's criterion In probability theory, Kolmogorov's criterion, named after Andrey Kolmogorov, is a theorem giving a necessary and sufficient condition for a Markov chain or continuous-time Markov chain to be stochastically identical to its time-reversed version. ...
states that the necessary and sufficient condition for a process to be reversible is that the product of transition rates around a closed loop must be the same in both directions.


Embedded Markov chain

One method of finding the stationary probability distribution, , of an
ergodic In mathematics, ergodicity expresses the idea that a point of a moving system, either a dynamical system or a stochastic process, will eventually visit all parts of the space that the system moves in, in a uniform and random sense. This implies tha ...
continuous-time Markov chain, ''Q'', is by first finding its embedded Markov chain (EMC). Strictly speaking, the EMC is a regular discrete-time Markov chain. Each element of the one-step transition probability matrix of the EMC, ''S'', is denoted by ''s''''ij'', and represents the
conditional probability In probability theory, conditional probability is a measure of the probability of an event occurring, given that another event (by assumption, presumption, assertion or evidence) has already occurred. This particular method relies on event B occu ...
of transitioning from state ''i'' into state ''j''. These conditional probabilities may be found by : s_ = \begin \frac & \text i \neq j \\ 0 & \text. \end From this, ''S'' may be written as :S = I - \left( \operatorname(Q) \right)^ Q where ''I'' is the
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. Terminology and notation The identity matrix is often denoted by I_n, or simply by I if the size is immaterial or ...
and diag(''Q'') is the
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagonal m ...
formed by selecting the
main diagonal In linear algebra, the main diagonal (sometimes principal diagonal, primary diagonal, leading diagonal, major diagonal, or good diagonal) of a matrix A is the list of entries a_ where i = j. All off-diagonal elements are zero in a diagonal matri ...
from the matrix ''Q'' and setting all other elements to zero. To find the stationary probability distribution vector, we must next find \varphi such that :\varphi S = \varphi, with \varphi being a row vector, such that all elements in \varphi are greater than 0 and \, \varphi\, _1 = 1. From this, may be found as :\pi = . (''S'' may be periodic, even if ''Q'' is not. Once is found, it must be normalized to a
unit vector In mathematics, a unit vector in a normed vector space is a vector (often a spatial vector) of length 1. A unit vector is often denoted by a lowercase letter with a circumflex, or "hat", as in \hat (pronounced "v-hat"). The term ''direction v ...
.) Another discrete-time process that may be derived from a continuous-time Markov chain is a δ-skeleton—the (discrete-time) Markov chain formed by observing ''X''(''t'') at intervals of δ units of time. The random variables ''X''(0), ''X''(δ), ''X''(2δ), ... give the sequence of states visited by the δ-skeleton.


See also

*
Kolmogorov equations (Markov jump process) In mathematics and statistics, in the context of Markov processes, the Kolmogorov equations, including Kolmogorov forward equations and Kolmogorov backward equations, are a pair of systems of differential equations that describe the time evolut ...


Notes


References

* * Leo Breiman (1992)
968 Year 968 ( CMLXVIII) was a leap year starting on Wednesday (link will display the full calendar) of the Julian calendar. Events By place Byzantine Empire * Emperor Nikephoros II receives a Bulgarian embassy led by Prince Boris ( ...
''Probability''. Original edition published by Addison-Wesley; reprinted by
Society for Industrial and Applied Mathematics Society for Industrial and Applied Mathematics (SIAM) is a professional society dedicated to applied mathematics, computational science, and data science through research, publications, and community. SIAM is the world's largest scientific soci ...
. (See Chapter 7) * * * J. L. Doob (1953) ''Stochastic Processes''. New York: John Wiley and Sons . * A. A. Markov (1971). "Extension of the limit theorems of probability theory to a sum of variables connected in a chain". reprinted in Appendix B of: R. Howard. ''Dynamic Probabilistic Systems, volume 1: Markov Chains''. John Wiley and Sons. * * S. P. Meyn and R. L. Tweedie (1993) ''Markov Chains and Stochastic Stability''. London: Springer-Verlag . online
MCSS
. Second edition to appear, Cambridge University Press, 2009. * Classical text. cf Chapter 6 ''Finite Markov Chains'' pp. 384ff. * John G. Kemeny &
J. Laurie Snell James Laurie Snell, often cited as J. Laurie Snell, (January 15, 1925 in Wheaton, Illinois – March 19, 2011 in Hanover, New Hampshire) was an American mathematician. Biography J. Laurie Snell was the son of Roy Snell, an adventure author, and ...
(1960) ''Finite Markov Chains'', D. van Nostrand Company * E. Nummelin. "General irreducible Markov chains and non-negative operators". Cambridge University Press, 1984, 2004. * Seneta, E. ''Non-negative matrices and Markov chains''. 2nd rev. ed., 1981, XVI, 288 p., Softcover Springer Series in Statistics. (Originally published by Allen & Unwin Ltd., London, 1973) * {{Queueing theory Markov processes