Continuous-time Markov Chains


Lecture 1

Continuous-time stochastic process \(\{X(t),t\geq 0\}\)

  • state: finite/countable values, {0,1,2, …}
  • time: continuous, \(t\geq 0\)
  • transition prob (time stationary/homogeneous), \(\Pr(X(t+s)=j|X(s)=i, X(u), 0\leq u\lt s) = \Pr(X(t)=j|X(0)=i)\).
  • \(T_i\) denote the amount of time that the process stays in state i before making a transition into a different state
    • \(\Pr(T_i>s+t|T_i>s)=\Pr(T_i>t)\)
      • the process is in state i at time s
      • by the Markovian property, that the prob that it remains in that state during the interval [s,s+t] is just the (unconditional) prob that it stays in state i for at least t time.
    • \(T_i\) is memoryless
    • \(T_i\) must follow the exponential dist

Alternative definition of continuous-time MC

  • a stochastic process that moves from state to state in accordance with a (discrete-time) MC
    • \(P_{ii}=0, \sum_{j\neq i} P_{ij}=1\)
  • the amount of time it spends in state i, before proceeding to the next state, is exponentially distributed with rate \(v_i\)
  • the amount of time the process spends in state i, and the next state visited, must be independent random variables

Memoryless continuous random variable X

  • \(\Pr(X>t+s|X>s)=\Pr(X>t)\) <==> \(\Pr(X>t+s)=\Pr(X>t)\Pr(X>s)\)
  • survival function of X: \(G(t)=\Pr(X>t)\) (right continuous)
  • G(t+s)=G(t)G(s)
    • easy to check \(G(nt)=G(t)^n\), or \(G(t)=G(nt)^{1/n}\).
    • so \(G(1/p)=G(1)^{1/p}, G(q/p)=G(1/p)^q=G(1)^{q/p}\) for any integer p,q.
    • rational number can approximate any real number, so \(G(t)=G(1)^t=\exp(-\log(1/G(1))t)\), an exponential dist with rate \(\lambda=\log(1/G(1))\).
  • geometric dist is the discrete analog: survival prob \((1-p)^k\).
  • exponential dist with rate \(\lambda\): \(f(t)=\lambda e^{-\lambda t}, G(t)=e^{-\lambda t}\); mean \(E[T]=1/\lambda\), variance \(Var[T]=1/\lambda^2\).
  • minimum of independent exp is still exp (rate is the sum of individual rates).
    • \(\Pr(\min(X,Y)>t)=\Pr(X>t)\Pr(Y>t)=e^{-t\lambda_x-t\lambda_y}\)
      • \(\Pr(\min_iY_i>t)=\prod_i\Pr(Y_i>t)=e^{-t\sum_i\lambda_i^y}\)
    • Pr(an exp is the minimum among a set of independent exps) \(\propto\) its rate
      • \(\Pr(X\lt Y)=E[E[I(X\lt Y)|X]]=E[e^{-X\lambda_y}]=\lambda_x/(\lambda_x+\lambda_y)\)
      • \(\Pr(X\lt Y_1,\cdots,X\lt Y_m)=E[E[I(X\lt Y_i,\forall i)|X]]\)
        • = \(E[e^{-X\sum_i\lambda_{i}^y}]=\lambda_x/(\lambda_x+\sum_i\lambda_i^y)\)
  • sum of iid exp is gamma(n,\(\lambda\)).
  • Failure/hazard rate
    • r(t)=f(t)/G(t)
    • \(\Pr(X\in (t,t+dt)|X>t)=r(t)dt\).
    • \(G(t) = \exp(-\int_0^t r(t)dt)\).
    • \(r(t)=\lambda\) for exponential dist.

Birth and Death Processes

  • Birth and death process
    • a system whose state at any time is represented by the number of people in the system at that time
    • whenever there are n people in the system
      • new arrivals enter the system at an exponential rate \(\lambda_n\) (birth/arrival rate)
      • people leave the system at an exponential rate \(\mu_n\) (departure/death rate)
        • time until the next arrival is exponentially distributed with mean \(1/\lambda_n\),
        • independent of the time until the next departure, which is itself exponentially distributed with mean \(1/\mu_n\).
    • a continuous-time MC with states {0,1,…} for which transitions from state n may go only to either state n-1 or state n+1
      • \(\nu_0=\lambda_0, \nu_i=\lambda_i+\mu_i, i>0\).
      • \(P_{01}=1,P_{i,i+1}=\lambda_i/(\lambda_i+\mu_i), P_{i,i-1}=\mu_i/(\lambda_i+\mu_i), i>0\)
    • Poisson process with rate \(\lambda\)
      • \(\mu_n=0,\lambda_n=\lambda\)
    • Sampling Poisson process \(N(t)\) (rate \(\lambda\)) with prob \((p,1-p)\)
      • leading to two independent Poisson processes: \(N_1(t)\) with rate \(\lambda p\) and \(N_2(t)\) with rate \(\lambda(1-p)\)
      • \(X(t)=N_1(t)-N_2(t)\) can be treated as a birth and death process
        • constant birth rate \(\lambda_n=\lambda p\) and death rate \(\mu_n=\lambda (1-p)\)
        • transition matrix \(P_{i,i+1}=p, P_{i,i-1}=1-p\)
  • Linear growth model with immigration (Ex 6.4, Page 361)
    • \(\mu_n=n\mu, n\geq 1; \lambda_n=n\lambda+\theta, n\geq 0\).
      • each member independently gives birth with rate \(\lambda\), and dies with rate \(\mu\)
      • external source of immigration with constant rate \(\theta\)
    • Deriving M(t)=E[X(t)]
      • Considering M(t+h)=E[X(t+h)]=E[E[X(t+h)|X(t)]]
      • for small h, 3 values for X(t+h) conditional on X(t)
        • X(t)+1, prob=\([\theta+X(t)\lambda]h+o(h)\)
        • X(t)-1, prob=\(X(t)\mu h+o(h)\)
        • X(t), prob = 1- the first two
      • so \(E[X(t+h)|X(t)]=X(t)+[\theta+X(t)\lambda-X(t)\mu]h + o(h)\)
        • expectation: \(M(t+h)=M(t)+(\lambda-\mu)M(t)h+\theta h+o(h)\)
        • \((M(t+h)-M(t))/h = (\lambda-\mu)M(t)+\theta+o(h)/h\)
        • as \(h\to 0\), \(M'(t)=(\lambda-\mu)M(t)+\theta := h(t)\)
        • \(h'(t)=(\lambda-\mu)M'(t)\) and \(h'(t)=(\lambda-\mu)h(t)\)
        • \(h(t) = Ke^{(\lambda-\mu)t}\) and \(\theta+(\lambda-\mu)M(t)=Ke^{(\lambda-\mu)t}\)
      • note \(M(0)=i\)
        • \(\lambda\neq\mu\): \(M(t)=\theta/(\lambda-\mu)[e^{(\lambda-\mu)t}-1]+ie^{(\lambda-\mu)t}\)
        • \(\lambda=\mu\): \(M(t)=\theta t+i\)
  • General birth-death process with birth rate \(\lambda_n\) and death rate \(\mu_n\) with \(\mu_0=0\)
    • \(T_i\) denote the time, starting from state i, it takes for the process to enter state i+1, \(i\geq 0\)
    • \(X_i\): time until the transition from i occurs: Exp dist with rate \(\lambda_i+\mu_i\)
    • Deriving \(E[T_i]\)
      • note \(E[T_0]=1/\lambda_0\) since \(\nu_0=\lambda_0\).
      • Define
        • \(I_i=1\) if the first transition from i is to i+1
        • \(I_i=0\) if the first transition from i is to i-1
        • \(\Pr(I_i=1)=\Pr(\mbox{birth before death})=\lambda_i/(\lambda_i+\mu_i)\)
      • independent of whether the first transition is from a birth or death, the time until it occurs is exponential with rate \(\lambda_i+\mu_i\)
      • \(E[T_i|I_i=1] = 1/(\lambda_i+\mu_i)\)
        • if the first transition is a birth, then the population size is at i+1, so no additional time is needed
        • \(E[T_i|I_i=1]=E[X_i|I_i=1] = E[X_i]\)
          • time until transition and next state visited are independent!
        • => \(Var(T_i|I_i=1)=1/(\lambda_i+\mu_i)^2\)
      • \(E[T_i|I_i=0] = 1/(\lambda_i+\mu_i) + E[T_{i-1}] +E[T_i]\)
        • if the first event is death, then the population size becomes i-1,
        • the additional time needed to reach i+1 is equal to the time it takes to return to state i (mean \(E[T_{i-1}]\))
        • plus the additional time it then takes to reach i+1 (mean \(E[T_i]\))
        • => \(Var(T_i|I_i=0)=1/(\lambda_i+\mu_i)^2+Var(T_{i-1})+Var(T_i)\)
      • ==> \(E[T_i] = 1/(\lambda_i+\mu_i)+\mu_i/(\lambda_i+\mu_i)(E[T_{i-1}]+E[T_i])\)
        • \(E[T_i]=1/\lambda_i+\mu_i/\lambda_iE[T_{i-1}]\)
    • expected time to go from state i to state j where i<j
      • \(E[T_i]+E[T_{i+1}]+\cdots+E[T_{j-1}]\)
    • For \(\lambda_n=\lambda, \mu_n=\mu\)
      • \(E[T_0]=1/\lambda\)
      • \(E[T_i]=1/\lambda+\mu/\lambda E[T_{i-1}]=(1+\mu E[T_{i-1}])/\lambda\)
      • \(E[T_1]=(1+\mu/\lambda)/\lambda, E[T_2] = (1+\mu/\lambda+[\mu/\lambda]^2)/\lambda\)
        • \(\lambda\neq \mu\): \(E[T_i] = (1-(\mu/\lambda)^{i+1})/(\lambda-\mu), i\geq 0\).
        • \(\lambda=\mu\): \(E[T_i] = (i+1)/\lambda\)
      • E[time to go from k to j]=\(\sum_{i=k}^{j-1}E[T_i]\), k<j.
        • \(\lambda\neq \mu\): \((j-k)/(\lambda-\mu)-(\mu/\lambda)^{k+1}/(\lambda-\mu)[1-(\mu/\lambda)^{j-k}]/(1-\mu/\lambda)\)
        • \(\lambda=\mu\): \((j(j+1)-k(k+1)/2/\lambda\)
    • Deriving \(Var[T_i]\)
      • Note \(E[T_i|I_i]=1/(\lambda_i+\mu_i)+(1-I_i)(E[T_{i-1}]+E[T_i])\)
        • \(I_i\) ~ Bernoulli(\(\lambda_i/(\lambda_i+\mu_i)\))
        • => \(Var(E[T_i|I_i])=(E[T_{i-1}]+E[T_i])^2Var(I_i)=(E[T_{i-1}]+E[T_i])^2\mu_i\lambda_i/(\mu_i+\lambda_i)^2\)
      • \(Var(T_i|I_i)=1/(\lambda_i+\mu_i)^2+(1-I_i)[Var(T_{i-1})+Var(T_i)]\)
        • => \(E[Var(T_i|I_i)]=1/(\lambda_i+\mu_i)^2+\mu_i\lambda_i/(\mu_i+\lambda_i)^2[Var(T_{i-1})+Var(T_i)]\)
      • ==> \(Var(T_i)=1/(\lambda_i(\lambda_i+\mu_i))+\frac{\mu_i}{\lambda_i}Var(T_{i-1})+\frac{\mu_i}{\mu_i+\lambda_i}(E[T_{i-1}]+E[T_i])^2\)
      • note \(Var(T_0)=1/\lambda_0^2\)
      • Var[time to go from k to j]=\(\sum_{i=k}^{j-1}Var[T_i]\), k<j.

Lecture 2

Negative binomial dist: NB(r,p)

  • number of successes in a sequence of iid Bernoulli trials before a specified number of failures occurs
  • \(\Pr(X=k) = \binom{k+r-1}{k} p^k(1-p)^r\)
    • \(r=1\): geometric dist (memoryless)
    • sum of iid geometric dist rvs is negative binomial dist
  • here binomial coeff \(\binom{k+r-1}{k}=\frac{(k+r-1)!}{k!(r-1)!}=\frac{(k+r-1)(k+r-2)\cdots(r)}{k!}\)
    • generally \(\Gamma(n)=(n-1)!\), so also \(\frac{\Gamma(k+r)}{\Gamma(r)\Gamma(k+1)}\)
    • it can also be written as \((-1)^k\frac{(-r)(-r-1)\cdots(-r-k+1)}{k!}=(-1)^k\binom{-r}{k}\) (negative binomial)
  • \(E[X+r]=r/(1-p), E[X]=rp/(1-p)\)
    • \(1/(1-p)\): average number of trials needed to have one failure
  • \(Var(X)=rp/(1-p)^2\)
    • in general \(Var(X)>E(X)\) (over-dispersion)
  • Consider \(r\to\infty\) and \(rp/(1-p)\to \lambda\), i.e., \(p=\frac{\lambda}{r+\lambda}\)
    • density is \(\frac{\Gamma(r+k)}{\Gamma(r)}\frac{\lambda^k}{k!}\frac{1}{(1+\lambda/r)^r}\)
    • as \(r\to\infty\), \((1+\lambda/r)^r\to e\), and \(\Gamma(r+k)/\Gamma(r)\to 1\)
    • so \(\Pr(X=k)\to \frac{\lambda^k}{k!} e^{-\lambda}\), i.e., Possion dist with mean \(\lambda\)
    • \(NB(r,\lambda/(\lambda+r))\to \mbox{Poisson}(\lambda)\) as \(r\to\infty\)
  • Poisson-Gamma mixture: can be treated as a mixture of Poisson dist with Gamma dist mixing
    • Poisson mean heterigeneity hence over-dispersion
    • \(Y\sim\mbox{Poisson}(\lambda)\)
      • \(\Pr(Y=k)\sim e^{-\lambda}\lambda^k/k!\)
      • \(E(Y|\lambda)=Var(Y|\lambda)=\lambda\)
    • \(\lambda\sim\mbox{Gamma}(r,p/(1-p))\) (shape \(r\) and scale \(p/(1-p)\))
      • \(f(\lambda) = \lambda^{r-1}\frac{e^{-\lambda(1-p)/p}}{\Gamma(r)(p/(1-p))^r}\)
      • \(E(\lambda)=rp/(1-p), Var(\lambda)=rp^2/(1-p)^2\)

Transition Probability Function

  • transition probabilities of a continuous-time MC
    • \(P_{ij}(t)=\Pr[X(t+s)=j|X(s)=i]\)

Pure birth process

  • Suppose that the process is presently in state i, and let j > i
    • \(X_k\) denote the time the process spends in state k before making a transition into state k+1
    • \(\sum_{k=i}^{j-1}X_k\) is the time it takes until the process enters state \(j\)
    • \(X(t)\lt j \leftrightarrow \sum_{k=i}^{j-1}X_k>t\)
    • \(\Pr[X(t)\lt j|X(0)=i]=\Pr[\sum_{k=i}^{j-1}X_k>t]\)
      • \(X_k\) are independent exps with rate \(\lambda_k\)
      • sum of independent exps with distinct rates (hypoexponential dist)
      • = \(\sum_{k=i}^{j-1}e^{-\lambda_kt}\prod_{r\neq k,r=i}^{j-1}\frac{\lambda_r}{\lambda_r-\lambda_k}\) (assuming \(\lambda_i\neq\lambda_j, \forall i\neq j\))
    • \(\Pr[X(t)=j|X(0)=i]=\Pr[X(t)\lt j+1|X(0)=i]-\Pr[X(t)\lt j|X(0)=i]\)
    • \(P_{ii}(t)=\Pr(X_i>t)=e^{-\lambda_it}\) (Exp dist surv prob)
  • When birth rate is linear: \(\lambda_n=n\lambda\)
    • \(P_{1j}(t)=e^{-\lambda t}(1-e^{-\lambda t})^{j-1}\) (geometric dist)
      • number of successes in a sequence of iid Bernoulli trials before a failure occurs
      • derivation: …
    • \(P_{ij}(t)=\binom{j-1}{i-1} e^{-i\lambda t}(1-e^{-\lambda t})^{j-1}\) (negative binomial dist)
      • each individual starts its own independent birth process.
      • sum of \(i\) independent geometric rvs

Lecture 3

Instantaneous transition rates: \(q_{ij}=\nu_iP_{ij}\)

  • \(q_{ij}\) is the rate, when in state i, at which the process makes a transition into state j
  • \(\nu_i=\sum_jq_{ij}\), \(P_{ij}=q_{ij}/\sum_jq_{ij}\)
  • birth-death process
    • birth rate \(\lambda_i\) and death rate \(\mu_i\)
    • \(\nu_i=\lambda_i+\mu_i\), \(P_{i,i+1}=\lambda_i/(\lambda_i+\mu_i)\)
    • \(q_{i,i+1}=\lambda_i\) (birth-rate), \(q_{i,i-1}=\mu_i\) (death-rate)
  • Lemma 6.2 \(\lim_{h\to 0}\frac{1-P_{ii}(h)}{h}=\nu_i\), \(\lim_{h\to 0}\frac{P_{ij}(h)}{h}=q_{ij}, i\neq j\)
    • \(\nu_i\) is the hazard rate of exp dist
    • \(1-P_{ii}(h)\): the probability that a process in state i at time 0 will not be in state i at time h, equals the probability that a transition occurs within time h plus something small compared to h
      • \(1-P_{ii}(h)=\nu_i h+o(h)\)
    • \(P_{ij}(h)\): the probability that the process goes from state i to state j in a time h, equals the probability that a transition occurs in this time multiplied by the probability that the transition is into state j plus something small compared to h
      • \(P_{ij}(h)=h\nu_iP_{ij}+o(h)\)
  • Chapman–Kolmogorov equations: \(P_{ij}(t+s)=\sum_{k=0}^{\infty}P_{ik}(t)P_{kj}(s), \forall s\geq 0, t\geq 0\)
    • Condition on \(X(t)=k\): \(P_{ij}(t+s)=\Pr[X(t+s)=j|X(0)=i]=\sum_k \Pr[X(t+s)=j, X(t)=k|X(0)=i]\)
    • In matrix notation: \(P(t+s)=P(t)P(s)\) (for discrete MC: \(P^{(n+m)}=P^{(n)}.P^{(m)}\))
    • Kolmogorov’s Backward Equations: \(P'_{ij}(t)=\sum_{k\neq i} q_{ik}P_{kj}(t)-v_iP_{ij}(t)\)
      • \(P_{ij}(h+t)-P_{ij}(t)=\sum_{k\neq i}P_{ik}(h)P_{kj}(t)-[1-P_{ii}(h)]P_{ij}(t)\)
      • taking limit \(h\to 0\) and interchange limit and summation
    • Kolmogorov’s Forward Equations: \(P'_{ij}(t)=\sum_{k\neq j} q_{kj}P_{ik}(t)-v_jP_{ij}(t)\)
      • \(P_{ij}(t+h)-P_{ij}(t)=\sum_{k\neq j}P_{ik}(t)P_{kj}(h)-[1-P_{jj}(h)]P_{ij}(t)\)
      • taking limit \(h\to 0\) and interchange limit and summation (under suitable regularity conditions)
  • Pure birth process
    • Backward equation
      • \(P_{ij}'(t)=\lambda_i[P_{i+1,j}(t)−P_{ij}(t)]\)
    • Forward equation
      • \(P_{ii}'(t)=-\lambda_iP_{ii}(t)\)
        • \(P_{ii}(t)=e^{-\lambda_i t}\)
      • \(P_{ij}'(t)=\lambda_{j-1}P_{i,j-1}(t)-\lambda_jP_{ij}(t)\)
        • \(P_{ij}(t)=\lambda_{j-1}e^{-\lambda_j t}\int_0^t e^{\lambda_j s}P_{i,j-1}(s)ds\)
        • treating \(e^{\lambda_j t}P_{ij}(t)\) as one function
  • Birth and death process
    • Backward equation
      • \(P_{0j}'(t)=\lambda_0[P_{1j}(t)−P_{0j}(t)]\)
      • \(P_{ij}'(t)=\lambda_iP_{i+1,j}(t)+\mu_iP_{i−1,j}(t)−(\lambda_i+\mu_i)P_{ij}(t)\)
    • Forward equation
      • \(P_{i0}'(t)=\mu_1P_{i1}(t)−\lambda_0P_{i0}(t)\)
      • \(P_{ij}'(t)=\lambda_{j-1}P_{i,j-1}(t)+\mu_{j+1}P_{i,j+1}(t)−(\lambda_j+\mu_j)P_{ij}(t)\)

Continuous-Time Markov Chain Consisting of Two States (Example 6.11, Page 371)

  • A machine works for an exponential amount of time having mean \(1/\lambda\) before breaking down; and it takes an exponential amount of time having mean \(1/\mu\) to repair the machine
    • two states: 1 (not working, being repaired), 0 (working)
    • A birth-death process
      • birth rate: \(\lambda_0=\lambda\) (0->1, rate of breaking down when it is working), \(\lambda_1=0\) (can only go from 1->0)
      • death rate: \(\mu_1=\mu\) (1->0, rate of being repaired), \(\mu_0=0\) (can only go from 0->1)
    • \(P_{00}(0)=1, P_{10}(0)=0\)
    • Differential equations
      • Backward equation
        • \(P_{0j}'(t)=\lambda_0[P_{1j}(t)−P_{0j}(t)]\)
        • j=0: \(P_{00}'(t)=\lambda [P_{10}(t)−P_{00}(t)]\)
      • Forward equation
        • \(P_{i0}'(t)=\mu_1P_{i1}(t)−\lambda_0P_{i0}(t)\)
        • i=0: \(P_{00}'(t)=\mu P_{01}(t)−\lambda P_{00}(t)\)
      • hence \(\lambda P_{10}(t)=\mu P_{01}(t)\)
        • so \(P_{00}'(t)=\mu P_{01}(t)− \lambda P_{00}(t) = \mu - (\mu+\lambda) P_{00}(t)\)
        • consider function \(h(t)=P_{00}(t)-\mu/(\mu+\lambda)\)
        • we have \(h'(t)=-(\mu+\lambda)h(t)\) and hence \(h(t)=Ke^{-(\mu+\lambda)t}\)
        • \(P_{00}(t)=Ke^{-(\mu+\lambda)t}+\mu/(\mu+\lambda)\)
        • \(P_{00}(t)=1\), so \(K=\lambda/(\mu+\lambda)\)
      • \(P_{00}(t)=\frac{\lambda}{\lambda+\mu}e^{-(\lambda+\mu)t}+\frac{\mu}{\lambda+\mu}\)
      • \(P_{10}(t)=\mu/\lambda[1-P_{00}(t)]=-\frac{\mu}{\lambda+\mu}e^{-(\lambda+\mu)t}+\frac{\mu}{\lambda+\mu}\)
    • some results
      • continuous-time MC: \(\nu_0=\lambda, \nu_1=\mu\). \(\lim_{t\to 0} P_{10}(t)/t=\mu, \lim_{t\to 0}P_{01}(t)/t=\lambda\)
      • embeded discrete MC: transition prob \(P_{10}=P_{01}=1\); stationary dist \(\pi_0=\pi_1=0.5\)
      • limiting prob of continuous-time MC: \(\lim_{t\to\infty}P_{10}(t)=\mu/(\mu+\lambda)=P_0\)

Lecture 4

Limiting Probabilities

  • \(P_{ij}(t)\) often converges to a limiting value that is independent of the initial state \(i\).
    • let \(P_j=\lim_{t\to\infty}P_{ij}(t)\), \(\sum_jP_j=1\) (stationary probabilities)
      • \(P_{ij}(t)\) will be constant if we start from the limiting prob.
    • Forward equation: \(P'_{ij}(t)=\sum_{k\neq j} q_{kj}P_{ik}(t)-\nu_jP_{ij}(t)\)
      • \(q_{kj}=v_kP_{kj}\)
      • \(\lim_{t\to\infty}P'_{ij}(t)=\sum_{k\neq j}q_{kj}P_k-\nu_jP_j\)
        • note \(\lim_{t\to\infty}P'_{ij}(t)=0\) (if not, \(P_{ij}(t)\) will go to infinity)
    • So \(\nu_jP_j=\sum_{k\neq j}q_{kj}P_k\) (balance equations)
      • Sufficient condition for \(P_j\) existing (ergodic chain)
        • all states of the MC communicate in the sense that starting in state i there is a positive probability of ever being in state j, for all i, j and
        • the MC is positive recurrent in the sense that, starting in any state, the mean time to return to that state is finite
      • \(P_j\) being the long-run proportion of time that the process is in state \(j\)
      • \(\nu_jP_j\) = rate at which the process leaves state j
        • when the process is in state j, it leaves at rate \(\nu_j\)
        • \(P_j\) is the proportion of time it is in state j
      • \(\sum_{k\neq j}q_{kj}P_k\) = rate at which the process enters state j
        • \(q_{kj}P_j\) = rate at which transitions from k to j occur
      • hence equality of the rates at which the process enters and leaves state j
  • Birth and death process
    • birth rate \(\lambda_n\), death rate \(\mu_n\)
      • embeded discrete-time MC: \(P_{n,n+1}=\lambda_n/(\lambda_n+\mu_n), P_{n,n-1}=\mu_n/(\lambda_n+\mu_n)\) (two competing Exp rvs)
      • rate \(v_n=\lambda_n+\mu_n\) (minimum of two Exp rvs)
      • \(q_{n,n+1}=\lambda_n, q_{n,n-1}=\mu_n\)
      • state 0: leaving rate \(\lambda_0P_0\); entering rate \(\mu_1P_1\)
      • state \(n\geq 1\): leaving rate \((\lambda_n+\mu_n)P_n\); entering rate \(\mu_{n+1}P_{n+1}+\lambda_{n-1}P_{n-1}\)
      • hence \((\lambda_n+\mu_n)P_n=\mu_{n+1}P_{n+1}+\lambda_{n-1}P_{n-1}\)
        • i.e., \(\lambda_nP_n - \lambda_{n-1}P_{n-1} = \mu_{n+1}P_{n+1} - \mu_nP_n\)
        • note for \(n=0\): \(\lambda_0P_0=\mu_1P_1\)
        • sum over 0,…,n: \(\lambda_nP_n=\mu_{n+1}P_{n+1}, n\geq 0\)
      • hence \(P_n=\frac{\prod_{i=1}^n\lambda_{i-1}}{\prod_{i=1}^n\mu_i}P_0\), and \(P_0=\frac{1}{1+\sum_{n=1}^{\infty}\frac{\prod_{i=1}^n\lambda_{i-1}}{\prod_{i=1}^n\mu_i}}\)
        • limiting prob exists iff \(\sum_{n=1}^{\infty}\frac{\prod_{i=1}^n\lambda_{i-1}}{\prod_{i=1}^n\mu_i}<\infty\)
    • linear growth model with immigration
      • \(\lambda_n=n\lambda+\theta, \mu_n=n\mu\)
      • previous condition reduces to \(\sum_{n=1}^{\infty}\frac{\prod_{i=1}^n(\theta+(i-1)\lambda)}{n!\mu^n}<\infty\)
        • consider the limit of ratio of two adjacent terms: \(\lim_n\frac{\theta+n\lambda}{(n+1)\mu}=\frac{\lambda}{\mu}\)
        • therefore when \(\lambda<\mu\), the limiting prob exists!
    • a birth-death process with \(\mu_n=\mu, n\geq 1\), \(\lambda_n=(M-n)\lambda, n\leq M\), \(\lambda_n=0, n>M\)
      • Continuous-Time Markov Chain Consisting of Finite States (Ex 6.13, Page 373)
      • \(P_n=\frac{(\lambda/\mu)^nM!/(M-n)!}{1+\sum_{n=1}^M(\lambda/\mu)^nM!/(M-n)!}, n=0,\ldots,M\)

Lecture 5

Time Reversibility

  • For a continuous-time ergodic MC with limiting \(P_j\)
    • consider its discrete-time transition matrix \(P_{ij}\)
      • its limiting prob \(\pi_i\): \(\pi_i=\sum_j\pi_jP_{ji}\)
    • Claim: \(P_i=\pi_i/\nu_i/(\sum_j\pi_j/\nu_j)\)
      • \(\nu_iP_i=\sum_{j\neq i} P_jq_{ji}\)
      • note \(P_{ii}=0\)
    • Going backward in time, the amount of time the process spends in state \(i\) is also exp with the same rate
      • assume the chain is already in limiting dist
      • \(\Pr(\mbox{process is in state i throughout } [t−s,t]|X(t)=i)\)
        • = \(\Pr(\mbox{process is in state i throughout } [t−s,t])/\Pr(X(t)=i)=\Pr(X(t-s)=i)e^{-\nu_is}/\Pr(X(t)=i)=e^{-\nu_is}\)
      • sequence of states visited by the reversed process constitutes a discrete-time MC with transition probabilities \(Q_{ij}=\pi_jP_{ji}/\pi_i\)
      • reversed process is a continuous-time MC with the same transition rates as the forward-time process and with one-stage transition probabilities \(Q_{ij}\)
    • Continuous-time MC is time reversible if \(\pi_iP_{ij}=\pi_jP_{ji}\)
      • time reversible, in the sense that the process reversed in time has the same probabilistic structure as the original process, when the embedded chain is time reversible
      • <=> \(P_iq_{ij}=P_jq_{ji}\)
      • \(P_i\) is the proportion of time in state i and \(q_{ij}\) is the rate when in state i that the process goes to j
      • the rate at which the process goes directly from state i to state j is equal to the rate at which it goes directly from j to i
  • Proposition 6.5 An ergodic birth and death process is time reversible
    • the rate at which a birth and death process goes from state i to state i+1 is equal to the rate at which it goes from i+1 to i
    • In any length of time t the number of transitions from i to i+1 must equal to within 1 the number from i+1 to i
      • since between each transition from i to i+1 the process must return to i, and this can only occur through i+1, and vice versa.
  • Proposition 6.7 If for some \(\{P_i\geq 0\}\), \(\sum_iP_i=1\), and \(P_iq_{ij}=P_jq_{ji}, \forall i\neq j\), then the continuous-time MC is time reversible and \(P_i\) is the limiting probability of being in state i.
    • \(\sum_{j\neq i}P_iq_{ij}=\sum_{j\neq i}P_jq_{ji}\)
    • note \(\nu_i=\sum_{j\neq i}q_{ij}\)
    • so \(v_iP_i=\sum_{j\neq i}P_jq_{ji}\) (balance equation)
  • Proposition 6.8 A time reversible chain with limiting prob \(P_j,j\in S\) that is truncated to the set \(A\in S\) and remains irreducible is also time reversible and has limiting probabilities \(P_j^A\) given by \(P_j^A=\frac{P_j}{\sum_{i\in A}P_i}, j\in A\).

Uniformization

  • Consider a continuous-time Markov chain with \(\nu_i=\nu, \forall i\)
    • N(t) denote the number of state transitions by time t, then \(\{N(t),t\geq 0\}\) will be a Poisson process with rate \(\nu\).
      • N(t) \(\sim\) Poisson(\(\nu t\))
    • \(P_{ij}(t)=\Pr(X(t)=j|X(0)=i)=\sum_{n=0}^{\infty}\Pr(N(t)=n|X(0)=i)\Pr(X(t)=j|X(0)=i,N(t)=n)\)
      • \(\Pr(N(t)=n|X(0)=i)=e^{-\nu t}(\nu t)^n/n!\) (Poisson rv)
      • \(\Pr(X(t)=j|X(0)=i,N(t)=n)=P_{ij}^n\)
        • knowing that N(t) = n gives us no information about which states were visited, since the distribution of time spent in each state is the same for all states
        • \(P_{ij}^n\) is the n-stage transition probability associated with the discrete-time MC with transition prob \(P_{ij}\)
    • So \(P_{ij}(t)=\sum_{n=0}^{\infty}P_{ij}^{n}e^{-\nu t}(\nu t)^n/n!\)
      • approximate \(P_{ij}(t)\) by taking a partial sum and then computing the n stage probabilities \(P_{ij}^n\).
  • uniformizing the rate in which a transition occurs from each state by introducing transitions from a state to itself
    • assume \(\nu_i\leq\nu, \forall i\)
    • Define \(P_{ij}^{\ast}\): \(1-\nu_i/\nu\) for \(j=i\) and \(\nu_i/\nu P_{ij}\) for \(j\neq i\).
    • \(P_{ij}(t)=\sum_{n=0}^{\infty}P_{ij}^{\ast n}e^{-\nu t}(\nu t)^n/n!\)

Computing the Transition Probabilities

  • Define \(r_{ij}\): \(q_{ij}\) if \(i\neq j\), and \(-\nu_i\) if \(i=j\).
    • Kolmogorov backward equations: \(P'_{ij}(t)=\sum_k r_{ik}P_{kj}(t)\)
    • Kolmogorov forward equations: \(P'_{ij}(t)=\sum_k r_{kj}P_{ik}(t)\)
  • Define corresponding matrix \(\textbf{R}, \textbf{P}(t), \textbf{P}'(t)\)
    • \(\textbf{P}'(t)=\textbf{R}\textbf{P}(t)=\textbf{P}(t)\textbf{R}\)
      • so \(\textbf{P}(t)=\textbf{P}(0)e^{\textbf{R}t}=e^{\textbf{R}t}\), since \(\textbf{P}(0)=\textbf{I}\).
      • \(e^{\textbf{R}t}=\sum_{n=0}^{\infty}\textbf{R}^nt^n/n!\)
  • Approximation of \(\textbf{P}(t)\)
    • \(e^{\textbf{R}t}=\lim_{n\to\infty}(\textbf{I}+\textbf{R}t/n)^n\)
      • taking \(n=2^k\): approximating by matrix square (multiplication)
      • for large n, \(\textbf{I}+\textbf{R}t/n\) all nonnegative elements
    • \(e^{-\textbf{R}t}=\lim_{n\to\infty}(\textbf{I}-\textbf{R}t/n)^n\)
      • \((\textbf{I}-\textbf{R}t/n)^{-1}\) nonnegative elements