Markov renewal process

From The Right Wiki
Revision as of 21:10, 12 July 2023 by imported>RowanElder (Copyediting)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigationJump to search

Markov renewal processes are a class of random processes in probability and statistics that generalize the class of Markov jump processes. Other classes of random processes, such as Markov chains and Poisson processes, can be derived as special cases among the class of Markov renewal processes, while Markov renewal processes are special cases among the more general class of renewal processes.

Definition

File:Marked point process.png
An illustration of a Markov renewal process

In the context of a jump process that takes states in a state space S, consider the set of random variables (Xn,Tn), where Tn represents the jump times and Xn represents the associated states in the sequence of states (see Figure). Let the sequence of inter-arrival times τn=TnTn1. In order for the sequence (Xn,Tn) to be considered a Markov renewal process the following condition should hold: Pr(τn+1t,Xn+1=j(X0,T0),(X1,T1),,(Xn=i,Tn))=Pr(τn+1t,Xn+1=jXn=i)n1,t0,i,jS

Relation to other stochastic processes

  1. Let Xn and Tn be as defined in the previous statement. Defining a new stochastic process Yt:=Xn for t[Tn,Tn+1), then the process Yt is called a semi-Markov process as it happens in a continuous-time Markov chain. The process is Markovian only at the specified jump instants, justifying the name semi-Markov.[1][2][3] (See also: hidden semi-Markov model.)
  2. A semi-Markov process (defined in the above bullet point) in which all the holding times are exponentially distributed is called a continuous-time Markov chain. In other words, if the inter-arrival times are exponentially distributed and if the waiting time in a state and the next state reached are independent, we have a continuous-time Markov chain.
    Pr(τn+1t,Xn+1=j(X0,T0),(X1,T1),,(Xn=i,Tn))=Pr(τn+1t,Xn+1=jXn=i)=Pr(Xn+1=jXn=i)(1eλit), for all n1,t0,i,jS,ij
  3. The sequence Xn in the Markov renewal process is a discrete-time Markov chain. In other words, if the time variables are ignored in the Markov renewal process equation, we end up with a discrete-time Markov chain.
    Pr(Xn+1=jX0,X1,,Xn=i)=Pr(Xn+1=jXn=i)n1,i,jS
  4. If the sequence of τs is independent and identically distributed, and if their distribution does not depend on the state Xn, then the process is a renewal. So, if the states are ignored and we have a chain of iid times, then we have a renewal process.
    Pr(τn+1tT0,T1,,Tn)=Pr(τn+1t)n1,t0

See also

References

  1. Medhi, J. (1982). Stochastic processes. New York: Wiley & Sons. ISBN 978-0-470-27000-4.
  2. Ross, Sheldon M. (1999). Stochastic processes (2nd ed.). New York [u.a.]: Routledge. ISBN 978-0-471-12062-9.
  3. Barbu, Vlad Stefan; Limnios, Nikolaos (2008). Semi-Markov chains and hidden semi-Markov models toward applications: their use in reliability and DNA analysis. New York: Springer. ISBN 978-0-387-73171-1.