HOME

TheInfoList



OR:

A continuous-time Markov chain (CTMC) is a continuous
stochastic process In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
in which, for each state, the process will change state according to an exponential random variable and then move to a different state as specified by the probabilities of a
stochastic matrix In mathematics, a stochastic matrix is a square matrix used to describe the transitions of a Markov chain. Each of its entries is a nonnegative real number representing a probability. It is also called a probability matrix, transition matrix, ''s ...
. An equivalent formulation describes the process as changing state according to the least value of a set of exponential random variables, one for each possible state it can move to, with the parameters determined by the current state. An example of a CTMC with three states \ is as follows: the process makes a transition after the amount of time specified by the holding time—an exponential random variable E_i, where ''i'' is its current state. Each random variable is independent and such that E_0\sim \text(6), E_1\sim \text(12) and E_2\sim \text(18). When a transition is to be made, the process moves according to the jump chain, a discrete-time Markov chain with stochastic matrix: :\begin 0 & \frac & \frac \\ \frac & 0 & \frac \\ \frac & \frac & 0 \end. Equivalently, by the property of competing exponentials, this CTMC changes state from state ''i'' according to the minimum of two random variables, which are independent and such that E_\sim \text(q_) for i\neq j where the parameters are given by the Q-matrix Q=(q_) :\begin -6 & 3 & 3 \\ 4 & -12 & 8 \\ 15 & 3 & -18 \end. Each non-diagonal entry q_ can be computed as the probability that the jump chain moves from state ''i'' to state ''j'', divided by the expected holding time of state ''i''. The diagonal entries are chosen so that each row sums to 0. A CTMC satisfies the
Markov property In probability theory and statistics, the term Markov property refers to the memoryless property of a stochastic process, which means that its future evolution is independent of its history. It is named after the Russian mathematician Andrey Ma ...
, that its behavior depends only on its current state and not on its past behavior, due to the memorylessness of the exponential distribution and of discrete-time Markov chains.


Definition

Let (\Omega,,\Pr) be a probability space, let S be a countable nonempty set, and let T=\mathbb R_ (T for "time"). Equip S with the
discrete metric Discrete may refer to: *Discrete particle or quantum in physics, for example in quantum theory *Discrete device, an electronic component with just one circuit element, either passive or active, other than an integrated circuit *Discrete group, a g ...
, so that we can make sense of right continuity of functions \mathbb R_\to S. A continuous-time Markov chain is defined by: * A
probability vector In mathematics and statistics, a probability vector or stochastic vector is a vector with non-negative entries that add up to one. The positions (indices) of a probability vector represent the possible outcomes of a discrete random variable, and ...
\lambda on S (which below we will interpret as the ''initial distribution'' of the Markov chain), and * A rate matrix Q on S, that is, a function Q:S^2\to\mathbb R such that # for all distinct i,j\in S, 0\le q_, # for all i\in S, \sum_q_=-q_. (Even if S is infinite, this sum is ''a priori'' well defined (possibly equalling +\infty) because each term appearing in the sum is nonnegative. ''A posteriori'', we know the sum must also be finite (not equal to +\infty), since we're assuming it's equal to -q_ and we've assumed Q is real valued. Some authors instead use a definition that's word-for-word the same except for a modified stipulation Q:S^2\to\mathbb R\cup\, and say Q is ''stable'' or ''totally stable'' to mean \operatornameQ\subseteq\mathbb R, i.e., every entry is real valued.) Note that the row sums of Q are 0: \forall i\in S,~\sum_q_=0, or more succinctly, Q\cdot1=0. This situation contrasts with the situation for discrete-time Markov chains, where all row sums of the transition matrix equal unity. Now, let X:T\to S^\Omega such that \forall t\in T~X(t) is (,(S))-measurable. There are three equivalent ways to define X being ''Markov with initial distribution \lambda and rate matrix Q'': via transition probabilities or via the jump chain and holding times. As a prelude to a transition-probability definition, we first motivate the definition of a ''regular'' rate matrix. We will use the transition-rate matrix Q to specify the dynamics of the Markov chain by means of generating a collection of ''transition matrices'' P(t) on S (t\in\mathbb R_), via the following theorem. We say Q is ''regular'' to mean that we do have uniqueness for the above system, i.e., that there exists exactly one solution. We say Q is ''irregular'' to mean Q is not regular. If S is finite, then there is exactly one solution, namely P=(e^)_, and hence Q is regular. Otherwise, S is infinite, and there exist irregular transition-rate matrices on S. If Q is regular, then for the unique solution P, for each t\in T, P(t) will be a
stochastic matrix In mathematics, a stochastic matrix is a square matrix used to describe the transitions of a Markov chain. Each of its entries is a nonnegative real number representing a probability. It is also called a probability matrix, transition matrix, ''s ...
. We will assume Q is regular from the beginning of the following subsection up through the end of this section, even though it is conventional to not include this assumption. (Note for the expert: thus we are not defining continuous-time Markov chains in general but only ''non-explosive'' continuous-time Markov chains.)


Transition-probability definition

Let P be the (unique) solution of the system (). (Uniqueness guaranteed by our assumption that Q is regular.) We say X is ''Markov with initial distribution \lambda and rate matrix Q'' to mean: for any nonnegative integer n\ge0, for all t_0,\dots,t_\in T such that t_0<\dots for all i_0,\dots,i_\in I, Using induction and the fact that \forall A,B\in~~\Pr(B)\ne0\rightarrow\Pr(A\cap B)=\Pr(A\mid B)\Pr(B), we can show the equivalence of the above statement containing () and the following statement: for all i\in I,~\Pr(X_0=i)=\lambda_i and for any nonnegative integer n\ge0, for all t_0,\dots,t_\in T such that t_0<\dots for all i_0,\dots,i_\in I such that 0<\Pr(X_0=i_0,\dots,X_=i_n) (it follows that 0<\Pr(X_=i_n)), It follows from continuity of the functions (P(t)_)_ (i,j\in S) that the trajectory (X_t(\omega))_ is almost surely right continuous (with respect to the
discrete metric Discrete may refer to: *Discrete particle or quantum in physics, for example in quantum theory *Discrete device, an electronic component with just one circuit element, either passive or active, other than an integrated circuit *Discrete group, a g ...
on S): there exists a \Pr-null set N such that \\subseteq N.


Jump-chain/holding-time definition


Sequences associated to a right-continuous function

Let f:T\to S be right continuous (when we equip S with the
discrete metric Discrete may refer to: *Discrete particle or quantum in physics, for example in quantum theory *Discrete device, an electronic component with just one circuit element, either passive or active, other than an integrated circuit *Discrete group, a g ...
). Define :h=h(f)=(\inf\)_)\cup\, let :H=H(f)=(h^0)_ be the ''holding-time sequence'' associated to f, choose s\in S, and let :y=y(f)=\left(\beginf(\sum_H_k)&\text\sum_H_k<+\infty,\\s&\text\end\right)_ be "the ''state sequence''" associated to f.


Definition of the jump matrix Π

The jump matrix \Pi, alternatively written \Pi(Q) if we wish to emphasize the dependence on Q, is the matrix \Pi=( =j_\cup\bigcup_(\\cup\), where Z=Z(Q)=\ is the ''zero set'' of the function (q_)_.


Jump-chain/holding-time property

We say X is ''Markov with initial distribution \lambda and rate matrix Q'' to mean: the trajectories of X are almost surely right continuous, let f be a modification of X to have (everywhere) right-continuous trajectories, \sum_H(f(\omega))_n=+\infty almost surely (note to experts: this condition says X is non-explosive), the state sequence y(f(\omega)) is a discrete-time Markov chain with initial distribution \lambda (jump-chain property) and transition matrix \Pi(Q), and \forall n\in\mathbb Z_~\forall B\in(\mathbb R_)~\Pr(H_n(f)\in B)=\operatorname(-q_)(B) (holding-time property).


Infinitesimal definition

We say X is ''Markov with initial distribution \lambda and rate matrix Q'' to mean: for all i\in S, \Pr(X(0)=i)=\lambda_i and for all i,j, for all t and for small strictly positive values of h, the following holds for all t\in T such that 0<\Pr(X(t)=i): :\Pr(X(t+h) = j \mid X(t) = i) = =j+ q_h + o(h), where the term =j/math> is 1 if i=j and otherwise 0, and the little-o term o(h) depends in a certain way on i,j,h. The above equation shows that q_ can be seen as measuring how quickly the transition from i to j happens for i\neq j, and how quickly the transition away from i happens for i= j.


Properties


Communicating classes

Communicating classes, transience, recurrence and positive and null recurrence are defined identically as for discrete-time Markov chains.


Transient behaviour

Write P(''t'') for the matrix with entries ''p''''ij'' = P(''X''''t'' = ''j'' ,  ''X''0 = ''i''). Then the matrix P(''t'') satisfies the forward equation, a
first-order differential equation In mathematics, an ordinary differential equation (ODE) is a differential equation (DE) dependent on only a single independent variable. As with any other DE, its unknown(s) consists of one (or more) function(s) and involves the derivatives ...
:P'(t) = P(t) Q, where the prime denotes differentiation with respect to ''t''. The solution to this equation is given by a
matrix exponential In mathematics, the matrix exponential is a matrix function on square matrix, square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exp ...
:P(t) = e^. In a simple case such as a CTMC on the state space . The general ''Q'' matrix for such a process is the following 2 × 2 matrix with ''α'',''β'' > 0 :Q = \begin -\alpha & \alpha \\ \beta & -\beta \end. The above relation for forward matrix can be solved explicitly in this case to give :P(t) = \begin \frac + \frace^ & \frac - \frace^ \\ \frac - \frace^ & \frac + \frace^ \end. Computing direct solutions is complicated in larger matrices. The fact that ''Q'' is the generator for a
semigroup In mathematics, a semigroup is an algebraic structure consisting of a set together with an associative internal binary operation on it. The binary operation of a semigroup is most often denoted multiplicatively (just notation, not necessarily th ...
of matrices :P(t+s) = e^ = e^ e^ = P(t) P(s) is used.


Stationary distribution

The stationary distribution for an irreducible recurrent CTMC is the probability distribution to which the process converges for large values of ''t''. Observe that for the two-state process considered earlier with P(''t'') given by :P(t) = \begin \frac + \frace^ & \frac - \frace^ \\ \frac - \frace^ & \frac + \frace^ \end, as ''t'' → ∞ the distribution tends to :P_\pi = \begin \frac & \frac \\ \frac & \frac \end. Observe that each row has the same distribution as this does not depend on starting state. The row vector ' may be found by solving :\pi Q = 0 with the constraint :\sum_ \pi_i = 1.


Example 1

The image to the right describes a continuous-time Markov chain with state-space and transition-rate matrix :: Q=\begin -0.025 & 0.02 & 0.005 \\ 0.3 & -0.5 & 0.2 \\ 0.02 & 0.4 & -0.42 \end. The stationary distribution of this chain can be found by solving \pi Q=0, subject to the constraint that elements must sum to 1 to obtain :\pi = \begin0.885 & 0.071 & 0.044 \end.


Example 2

The image to the right describes a discrete-time Markov chain modeling
Pac-Man ''Pac-Man,'' originally called in Japan, is a 1980 maze video game developed and published by Namco for arcades. In North America, the game was released by Midway Manufacturing as part of its licensing agreement with Namco America. The pla ...
with state-space . The player controls Pac-Man through a maze, eating pac-dots. Meanwhile, he is being hunted by ghosts. For convenience, the maze shall be a small 3x3-grid and the ghosts move randomly in horizontal and vertical directions. A secret passageway between states 2 and 8 can be used in both directions. Entries with probability zero are removed in the following transition-rate matrix: Q=\begin -1 & \frac & & \frac\\ \frac & -1 & \frac & & \frac&&&\frac\\ & \frac & -1 & & & \frac\\ \frac & & & -1 & \frac & & \frac\\ & \frac & & \frac & -1 & \frac & & \frac\\ & & \frac & & \frac & -1& & & \frac\\ & & & \frac & & & -1 & \frac\\ & \frac & && \frac & & \frac & -1 & \frac\\ & & & & & \frac & & \frac & -1\end This Markov chain is irreducible, because the ghosts can fly from every state to every state in a finite amount of time. Due to the secret passageway, the Markov chain is also aperiodic, because the ghosts can move from any state to any state both in an even and in an uneven number of state transitions. Therefore, a unique stationary distribution exists and can be found by solving \pi Q=0, subject to the constraint that elements must sum to 1. The solution of this linear equation subject to the constraint is \pi=(7.7,15.4,7.7,11.5,15.4,11.5,7.7,15.4,7.7)\%. The central state and the border states 2 and 8 of the adjacent secret passageway are visited most and the corner states are visited least.


Time reversal

For a CTMC ''X''''t'', the time-reversed process is defined to be \hat X_t = X_. By
Kelly's lemma In probability theory, Kelly's lemma states that for a stationary continuous-time Markov chain A continuous-time Markov chain (CTMC) is a continuous stochastic process in which, for each state, the process will change state according to an expon ...
this process has the same stationary distribution as the forward process. A chain is said to be reversible if the reversed process is the same as the forward process.
Kolmogorov's criterion In probability theory, Kolmogorov's criterion, named after Andrey Kolmogorov, is a theorem giving a necessary and sufficient condition for a Markov chain or continuous-time Markov chain to be stochastically identical to its time-reversed version. ...
states that the necessary and sufficient condition for a process to be reversible is that the product of transition rates around a closed loop must be the same in both directions.


Embedded Markov chain

One method of finding the
stationary probability distribution Stationary distribution may refer to: * and , a special distribution for a Markov chain such that if the chain starts with its stationary distribution, the marginal distribution of all states at any time will always be the stationary distribution. ...
, , of an
ergodic In mathematics, ergodicity expresses the idea that a point of a moving system, either a dynamical system or a stochastic process, will eventually visit all parts of the space that the system moves in, in a uniform and random sense. This implies th ...
continuous-time Markov chain, ''Q'', is by first finding its embedded Markov chain (EMC). Strictly speaking, the EMC is a regular discrete-time Markov chain. Each element of the one-step transition probability matrix of the EMC, ''S'', is denoted by ''s''''ij'', and represents the
conditional probability In probability theory, conditional probability is a measure of the probability of an Event (probability theory), event occurring, given that another event (by assumption, presumption, assertion or evidence) is already known to have occurred. This ...
of transitioning from state ''i'' into state ''j''. These conditional probabilities may be found by : s_ = \begin \frac & \text i \neq j,\\ 0 & \text. \end From this, ''S'' may be written as :S = I - \left( \operatorname(Q) \right)^ Q where ''I'' is the
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
and diag(''Q'') is the
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagon ...
formed by selecting the
main diagonal In linear algebra, the main diagonal (sometimes principal diagonal, primary diagonal, leading diagonal, major diagonal, or good diagonal) of a matrix A is the list of entries a_ where i = j. All off-diagonal elements are zero in a diagonal matrix ...
from the matrix ''Q'' and setting all other elements to zero. To find the stationary probability distribution vector, we must next find \varphi such that :\varphi S = \varphi, with \varphi being a row vector, such that all elements in \varphi are greater than 0 and \, \varphi\, _1 = 1. From this, may be found as :\pi = . (''S'' may be periodic, even if ''Q'' is not. Once is found, it must be normalized to a
unit vector In mathematics, a unit vector in a normed vector space is a Vector (mathematics and physics), vector (often a vector (geometry), spatial vector) of Norm (mathematics), length 1. A unit vector is often denoted by a lowercase letter with a circumfle ...
.) Another discrete-time process that may be derived from a continuous-time Markov chain is a δ-skeleton—the (discrete-time) Markov chain formed by observing ''X''(''t'') at intervals of δ units of time. The random variables ''X''(0), ''X''(δ), ''X''(2δ), ... give the sequence of states visited by the δ-skeleton.


See also

*
Kolmogorov equations (Markov jump process) In probability theory, Kolmogorov equations characterize continuous-time Markov processes. In particular, they describe how the probability of a continuous-time Markov process in a certain state changes over time. There are four distinct equatio ...


Notes


References

* * Leo Breiman (1992)
968 Year 968 ( CMLXVIII) was a leap year starting on Wednesday of the Julian calendar. Events By place Byzantine Empire * Emperor Nikephoros II receives a Bulgarian embassy led by Prince Boris (the son of Tsar Peter I of Bulgaria), wi ...
''Probability''. Original edition published by Addison-Wesley; reprinted by
Society for Industrial and Applied Mathematics Society for Industrial and Applied Mathematics (SIAM) is a professional society dedicated to applied mathematics, computational science, and data science through research, publications, and community. SIAM is the world's largest scientific soci ...
. (See Chapter 7) * * *
J. L. Doob Joseph Leo Doob (February 27, 1910 – June 7, 2004) was an American mathematician, specializing in Mathematical analysis, analysis and probability theory. The theory of Martingale (probability theory), martingales was developed by Doob. Early ...
(1953) ''Stochastic Processes''. New York: John Wiley and Sons . * A. A. Markov (1971). "Extension of the limit theorems of probability theory to a sum of variables connected in a chain". reprinted in Appendix B of: R. Howard. ''Dynamic Probabilistic Systems, volume 1: Markov Chains''. John Wiley and Sons. * * S. P. Meyn and R. L. Tweedie (1993) ''Markov Chains and Stochastic Stability''. London: Springer-Verlag . online
MCSS
. Second edition to appear, Cambridge University Press, 2009. * Classical text. cf Chapter 6 ''Finite Markov Chains'' pp. 384ff. * John G. Kemeny & J. Laurie Snell (1960) ''Finite Markov Chains'', D. van Nostrand Company * E. Nummelin. ''General irreducible Markov chains and non-negative operators''. Cambridge University Press, 1984, 2004. * * Seneta, E. ''Non-negative matrices and Markov chains''. 2nd rev. ed., 1981, XVI, 288 p., Softcover Springer Series in Statistics. (Originally published by Allen & Unwin Ltd., London, 1973) * {{Queueing theory Markov processes