Chapter 7: Problem 46
Consider a semi-Markov process in which the amount of time that the process spends in each state before making a transition into a different state is exponentially distributed. What kind of process is this?
/*! This file is auto-generated */ .wp-block-button__link{color:#fff;background-color:#32373c;border-radius:9999px;box-shadow:none;text-decoration:none;padding:calc(.667em + 2px) calc(1.333em + 2px);font-size:1.125em}.wp-block-file__button{background:#32373c;color:#fff;text-decoration:none}
Learning Materials
Features
Discover
Chapter 7: Problem 46
Consider a semi-Markov process in which the amount of time that the process spends in each state before making a transition into a different state is exponentially distributed. What kind of process is this?
All the tools & learning materials you need for study success - in one app.
Get started for free
Consider a renewal process with mean interarrival time \(\mu\). Suppose that each event of this process is independently "counted" with probability \(p .\) Let \(N_{C}(t)\) denote the number of counted events by time \(t, t>0\). (a) Is \(N_{C}(t), t \geqslant 0\) a renewal process? (b) What is \(\lim _{t \rightarrow \infty} N_{C}(t) / t ?\)
For a renewal process, let \(A(t)\) be the age at time \(t\). Prove that if \(\mu<\infty\), then with probability 1 $$ \frac{A(t)}{t} \rightarrow 0 \quad \text { as } t \rightarrow \infty $$
Let \(X_{1}, X_{2}, \ldots\) be a sequence of independent random variables. The nonnegative integer valued random variable \(N\) is said to be a stopping time for the sequence if the event \(\\{N=n\\}\) is independent of \(X_{n+1}, X_{n+2}, \ldots\). The idea being that the \(X_{i}\) are observed one at a time-first \(X_{1}\), then \(X_{2}\), and so on-and \(N\) represents the number observed when we stop. Hence, the event \(\\{N=n\\}\) corresponds to stopping after having observed \(X_{1}, \ldots, X_{n}\) and thus must be independent of the values of random variables yet to come, namely, \(X_{n+1}, X_{n+2}, \ldots\) (a) Let \(X_{1}, X_{2}, \ldots\) be independent with $$ P\left\\{X_{i}=1\right\\}=p=1-P\left\\{X_{i}=0\right\\}, \quad i \geqslant 1 $$ Define $$ \begin{aligned} &N_{1}=\min \left\\{n: X_{1}+\cdots+X_{n}=5\right\\} \\ &N_{2}=\left\\{\begin{array}{ll} 3, & \text { if } X_{1}=0 \\ 5, & \text { if } X_{1}=1 \end{array}\right. \\ &N_{3}=\left\\{\begin{array}{ll} 3, & \text { if } X_{4}=0 \\ 2, & \text { if } X_{4}=1 \end{array}\right. \end{aligned} $$ Which of the \(N_{i}\) are stopping times for the sequence \(X_{1}, \ldots ?\) An important result, known as Wald's equation states that if \(X_{1}, X_{2}, \ldots\) are independent and identically distributed and have a finite mean \(E(X)\), and if \(N\) is a stopping time for this sequence having a finite mean, then $$ E\left[\sum_{i=1}^{N} X_{i}\right]=E[N] E[X] $$ To prove Wald's equation, let us define the indicator variables \(I_{i}, i \geqslant 1\) by $$ I_{i}=\left\\{\begin{array}{ll} 1, & \text { if } i \leqslant N \\ 0, & \text { if } i>N \end{array}\right. $$ (b) Show that $$ \sum_{i=1}^{N} X_{i}=\sum_{i=1}^{\infty} X_{i} I_{i} $$ From part (b) we see that $$ \begin{aligned} E\left[\sum_{i=1}^{N} X_{i}\right] &=E\left[\sum_{i=1}^{\infty} X_{i} I_{i}\right] \\ &=\sum_{i=1}^{\infty} E\left[X_{i} I_{i}\right] \end{aligned} $$ where the last equality assumes that the expectation can be brought inside the summation (as indeed can be rigorously proven in this case). (c) Argue that \(X_{i}\) and \(I_{i}\) are independent. Hint: \(I_{i}\) equals 0 or 1 depending on whether or not we have yet stopped after observing which random variables? (d) From part (c) we have $$ E\left[\sum_{i=1}^{N} X_{i}\right]=\sum_{i=1}^{\infty} E[X] E\left[I_{i}\right] $$ Complete the proof of Wald's equation. (e) What does Wald's equation tell us about the stopping times in part (a)?
Wald's equation can be used as the basis of a proof of the elementary renewal
theorem. Let \(X_{1}, X_{2}, \ldots\) denote the interarrival times of a renewal
process and let \(N(t)\) be the number of renewals by time \(t\).
(a) Show that whereas \(N(t)\) is not a stopping time, \(N(t)+1\) is.
Hint: Note that
$$
N(t)=n \Leftrightarrow X_{1}+\cdots+X_{n} \leqslant t \quad \text { and }
\quad X_{1}+\cdots+X_{n+1}>t
$$
(b) Argue that
$$
E\left[\sum_{i=1}^{N(t)+1} X_{i}\right]=\mu[m(t)+1]
$$
(c) Suppose that the \(X_{i}\) are bounded random variables. That is, suppose
there is a constant \(M\) such that \(P\left\\{X_{i}
Compute the renewal function when the interarrival distribution \(F\) is such that $$ 1-F(t)=p e^{-\mu_{1 t}}+(1-p) e^{-\mu_{2} t} $$
What do you think about this solution?
We value your feedback to improve our textbook solutions.