0 for 1 _
Ym+l < 0,
j=l this implies that m
for all i,
L Pij Yj :0:: Yi - Ym+l > Yi j=l
and hence the impossibility that 'L/}=I Pij Yj > maxi {Yi}. It follows that statement (i) holds, which is to say that there exists a non-negative vector x = (xl, x2, ... , Xm) such that x(P - I) = 0 and I:t=l Xi = 1; such an xis the required eigenvector.
3. Thinking of Xn+ 1 as the amount you may be sure of winning, you seek a betting scheme x such that Xn+ 1 is maximized subject to the inequalities n
for 1 :S j :Sm.
Xn+l :S LXitij i=l
Writing aij = -tij for 1 :S i :S nand an+l,j = 1, we obtain the linear program: maximize
Xn+l
subject to
n+l L Xiaij :S 0 i=l
for 1 :S j :Sm.
The dua11inear program is: m
minimize
0
subject to
L aij Yj = 0 j=l
for 1 :S i :S n,
m
Lan+l,jYj = 1, j=l
Yj :0::0
for 1 :S j :Sm.
Re-expressing the aij in terms of the tij as above, the dual program takes the form: m
minimize
0
subject to
L tij Pj j=l
=0
for 1 :S i :S n,
m
L Pj = 1, j=l
Pj :::: 0
for 1 :S j :Sm.
The vector x = 0 is a feasible solution of the primal program. The dual program has a feasible solution if and only if statement (a) holds. Therefore, if (a) holds, the dual program has minimal value 0, whence by the duality theorem of linear programming, the maximal value of the primal program is 0, in contradiction of statement (b). On the other hand, if (a) does not hold, the dual has no feasible solution, and therefore the primal program has no optimal solution. That is, the objective function of the primal is unbounded, and therefore (b) holds. [This was proved by De Finetti in 1937.] 4. Use induction, the claim being evidently true when n = 1. Suppose it is true for n = m. Certainly pm+l is of the correct form, and the equation pm+ 1x' = P(Pmx') with x = (1, w, w2 ) yields in its first row
288
Solutions
Branching processes revisited
[6.6.5]-[6.7.2]
as required. 5. The first part follows from the fact that Jr1 1 = 1 if and only if 1rU = 1. The second part follows from the fact that ni > 0 for all i if P is finite and irreducible, since this implies the invertibility of 1-P+ U.
6. The chessboard corresponds to a graph with 8 x 8 = 64 vertices, pairs of which are connected by edges when the corresponding move is legitimate for the piece in question. By Exercises (6.4.6), (6.5.9), we need only check that the graph is connected, and to calculate the degree of a comer vertex. (a) For the king there are 4 vertices of degree 3, 24 of degree 5, 36 of degree 8. Hence, the number of edges is 210 and the degree of a comer is 3. Therefore tt(king) = 420/3 = 140. (b) tt(queen) = (28 x 21 + 20 x 23 + 12 x 25 + 4 x 27)/21 = 208/3. (c) We restrict ourselves to the set of 32 vertices accessible from a given comer. Then tt(bishop) = (14x7+10x9+6x 11+2x 13)/7=40. (d) tt(knight) = (4 X 2 + 8 X 3 + 20 X 4 + 16 X 6 + 16 X 8)/2 = 168. (e) tt(rook) = 64 x 14/14 = 64. 7. They are walking on a product space of 8 x 16 vertices. Of these, 6 x 16 have degree 6 x 3 and 16 x 2 have degree 6 x 5. Hence tt(C) = (6
X
16
X
6
X
3 + 16
X
2
X
6
X
5)/18 = 448/3.
IP- All =(A.- l)(A. + !)(A.+ ~). Tedious computation yields the eigenvectors, and thus
8.
0 0)
-4 -1
3
-3
5
-1 1
.
6. 7 Solutions. Branching processes revisited 1. We have (using Theorem (5.4.3), or the fact that Gn+1(s) = G(Gn(s))) that the probability generating function of Zn is n- (n- l)s Gn(s) = , n + 1- ns so that n )k-1 nk-1 ( n )k+1 IP(Zn = k) = n + 1 n+ 1 n+ 1 - (n + l)k+ 1
(n -1) (
fork :=::: 1. Therefore, for y > 0, as n --+ oo,
1 llD(Z <2yn1Z >0)j[
2.
n-
l2ynJ
- 1 - Gn(O)
n
"""' t:J
nk-1
(
(n + l)k+1
=1-
1 ) - L2ynJ 1+-n -+l-e- 2Y.
Using the independence of different lines of descent, JE(s
z
n
.
.
~ skiP(Zn = k, extinction)
I extmct10n) = L
k=O
IP( extinction)
where Gn is the probability generating function of Zn.
289
~ skiP(Zn = k)r/ 1 = L = -Gn(sry), k=O
TJ
TJ
[6.7.3]-[6.8.1] 3.
Markov chains
Solutions
We have that 17 = G(17). In this case G(s) = q(l - ps)- 1, and therefore 17 = q I p. Hence
P q{pn _ qn _ p(sqlp)(pn-1 _ qn-1)} 1 - G n (s 17) = - · ..::.-.::---,-,..-::---,-7--=..:._::_-"----=---'17 q pn+1 _ qn+1 _ p(sqlp)(pn _ qn) p{qn _ Pn _ qs(qn-1 _ Pn-1)} qn+1 _ pn+1 _ qs(qn _ pn) ' which is Gn (s) with p and q interchanged.
4.
(a) Using the fact that var(X I X > 0) ::=: 0, JE(X 2 ) = JE(X 2
IX
> O)JP(X > 0) :::: lE(X
IX
> 0) 2 JP(X > 0) = lE(X)lE(X
IX
> 0).
(b) Hence
lE(Z~)
n
2
lE(Znl t-t I Zn > 0) < = lE(Wn) - t-tnlE(Zn) where Wn = ZnllE(Zn). By an easy calculation (see Lemma (5.4.2)),
where 0" 2 = var(Z 1 ) = plq 2 . (c) Doing the calculation exactly, lE Z n ( nlt-t
Z > 0 - lE(Znlt-tn) 1 ) - IP'(Zn > 0)- 1- Gn(O)--+ 1-11
I n
where 17 =II"(ultimate extinction) = q I p.
6.8 Solutions. Birth processes and the Poisson process 1. Let F and W be the incoming Poisson processes, and let N(t) = F(t)+ W(t). Certainly N(O) = 0 and N is non-decreasing. Arrivals of flies during [0, s] are independent of arrivals during (s, t ], if s < t; similarly for wasps. Therefore the aggregated arrival process during [0, s] is independent of the aggregated process during (s, t]. Now IP'(N(t +h)=
n +II N(t)
=
n)
= JP(A Ll B)
where
A = {one fly arrives during (t, t + h l},
B = {one wasp arrives during (t, t + h l}.
We have that JP(A Ll B) = JP(A) + JP(B) - II"( A
n B)
= A.h + t-th- (A.h)(t-th) + o(h) = (A.+ t-t)h + o(h). Finally
IP'(N(t +h)> n +II N(t) =
n)::::: IP'(A n B)+ JP(C u D),
290
Solutions, [6.8.2]-[6.8.4]
Birth processes and the Poisson process
where C ={two or more flies arrive in (t, t + h]} and D ={two or more wasps arrive in (t, t + h]}. This probability is no greater than (A.h)(p,h) + o(h) = o(h). 2. Let I be the incoming Poisson process, and let G be the process of arrivals of green insects. Matters of independence are dealt with as above. Finally, IP'(G(t +h)= n +II G(t) =
n)
n)
= p!P'(I(t +h)= n +II I(t) =
+ o(h) = pA.h + o(h),
11"( G(t +h) > n +II G(t) = n) ::=:II" (I (t +h) > n +II I (t) = n) = o(h). 3.
Conditioning on T1 and using the time-homogeneity of the process,
1P'(E(t)>xiT1=u)=
{
IP'(E(t-u)>x)
if u :::: t,
~
iftt+x,
(draw a diagram to help you see this). Therefore IP'(E(t) > x) = fooo IP'(E(t) >xI T1 = u)A.e-Audu = t!P'(E(t-u) >x)A.e-Audu+1
Jo
00
A.e-Audu.
t+x
You may solve the integral equation using Laplace transforms. Alternately you may guess the answer and then check that it works. The answer is IP'(E(t) ::=: x) = I - e-h, the exponential distribution. Actually this answer is obvious since E(t) > x if and only if there is no arrival in [t, t + x ], an event having probability e -Ax. 4.
The forward equation is i :::: j,
with boundary conditions Pi} (0) = Oij, the Kronecker delta. We write Gi (s, t) = "E-1 si Pi} (t), the probability generating function of B(t) conditional on B(O) = i. Multiply through the differential equation by si and sum over j:
a partial differential equation with boundary condition Gi (s, 0) = si. This may be solved in the usual way to obtain Gi(s, t) = g(eAt(l- s- 1)) for some function g. Using the boundary condition, we find that g(I- s- 1) = si and so g(u) = (1 - u)-i, yielding I (se-At)i Gi(s, t) = {I- eAI(I- s-1 )}i = {I- s(l- e-At)}i.
The coefficient of si is, by the binomial series, j
as required.
291
?:_
i,
[6.8.5]-[ 6.8.6]
Solutions
Markov chains
Alternatively use induction. Set j =ito obtain pji(t) = -Aipu(t) (remember Pi,i-1 (t) = 0), and therefore Pi i (t) = e- Ai t. Rewrite the differential equation as d ( Pij(t)e A1't) = A(j -l)Pi,j-t(t)e A1't . dt
Set j = i + 1 and solve to obtain Pi,i+l (t) = ie-Ait ( 1- e-At). Hence(*) holds, by induction. The mean is E(B(t)) = !_GJ(S,
as
t)l
=/eM, s=l
by an easy calculation. Similarly var(B(t)) = A + E(B(t)) - E(B(t)) 2 where A= -a22 G1(s,t)
as
I
.
s=l
Alternatively, note that B(t) has the negative binomial distribution with parameters e-At and I.
5.
The forward equations are P~(t) = An-lPn-t(t)- AnPn(t),
n 2: 0,
where Ai = i A+ v. The process is honest, and therefore m(t) = L::n npn(t) satisfies 00
00
m'(t) = Ln[(n -l)A
+ v]Pn-t(f)-
n=l
Ln(nA + v)pn(t) n=O
00
= L
{ A[(n + l)n- n 2 ] + v[(n + 1)- nl} Pn(t)
n=O 00
=
L(An n=O
+ v)pn(t)
= Am(t)
+ v.
Solve subject to m(O) = 0 to obtain m(t) = v(eJ..t- 1)/A.
6. Using the fact that the time to the nth arrival is the sum of exponential interarrival times (or using equation (6.8.15)), we have that
is given by
which may be expressed, using partial fractions, as
where n
a·z-
A.
IT-1A·-A·
j=O 1
Hi
292
z
Solutions
Continuous-time Markov chains
[6.8.7]-[6.9.1]
so long as Ai =!= A.J whenever i =!= j. The Laplace transform Pn may now be inverted as
See also Exercise (4.8.4).
7. Let Tn be the time of the nth arrival, and letT= Exercise (6.8.6), AnPn(())
limn~oo
Tn = sup{t: N(t) < oo}. Now, as in
. =ITn __). z= lE(e-eTn)
+ () Xo +X 1 + · · ·+ Xn where Xk is the (k + l)th interarrival time, a random variable which is i=O Ai
since Tn = exponentially distributed with parameter Ak· Using the continuity theorem, lE(e-eTn) -+ 1E(e-eT) as n-+ oo, whence AnPn(()) -+ 1E(e-eT) as n-+ oo, which may be inverted to obtain AnPn(t) -+ f(t) as n-+ oo where f is the density function ofT. Now
I
lE(N(t) N(t) < oo) = L:=;onPn(t) L:=n=O Pn (t) which converges or diverges according to whether or not L:=n npn (t) converges. However Pn (t) ~ A.; 1 f (t) as n -+ oo, so that L:=n npn (t) < oo if and only if L:=n nA.; 1 < oo. When A.n = (n
+ ~) 2 , we have that lE(e-eT) =
IT 00
{
n=O
1+
() 1 2 (n + 2)
}-1
= sech
(nv'e).
Inverting the Laplace transform (or consulting a table of such transforms) we find that
where 81 is the first Jacobi theta function.
6.9 Solutions. Continuous-time Markov chains 1.
(a) We have that
where P12 = 1 - Plio P21 = 1 - P22· Solve these subject to Pi} (t) = 8ij, the Kronecker delta, to obtain that the matrix P1 = (Pi} (t)) is given by
(b) There are many ways of calculating Gn; let us use generating functions. Note first that G0 =I, the identity matrix. Write
n ::=: 0,
293
[6.9.2]-[6.9.4]
Solutions
Markov chnins
and use the equation Gn+ 1 = G · Gn to find that
Hence an+1 = -(pfA.)cn+1 for n 2: 0, and the first difference equation becomes an+1 = -(A.+ p,)an, n 2: 1, which, subject to a1 = -p,, has solution an = (-l)np,(A. + p,)n- 1, n 2: 1. Therefore Cn = (-l)n+ 1A.(A. + p,)n- 1 for n 2: 1, and one may see similarly that bn =-an, dn = -en for n 2: 1. Using the facts that ao =do = 1 and bo = co = 0, we deduce that L:~0 (tn jn!)Gn = Pt where P 1 is given in part (a). (c) With1r = (n1, n2), we have that -p,n1 + A.n2 = 0 and p,n1- A.n2 = 0, whence n1 = (A.jp,)n2. In addition, n1 + n2 = 1 if n1 = A./(A. + p,) = 1- n2. 2.
(a) The required probability is lP'(X(t) = 2, X(3t) = 1 I X(O) = 1P'(X(3t) = 1 I X(O) = 1)
1)
p!2(t)P21 (2t)
Pll (3t)
using the Markov property and the homogeneity of the process. (b) Likewise, the required probability is p!2(t)p21 (2t)pu (t)
Pu (3t)pu (t)
the same as in part (a).
3. The interarrival times and runtimes are independent and exponentially distributed. It is the lackof-memory property which guarantees that X has the Markov property. The state space isS= {0, 1, 2, ... } and the generator is
G~
c ~
A. -(A.+ p,) p,
0 A. -(A.+ p,)
0 0 A.
...
)
Solutions of the equation 1rG = 0 satisfy -A.no + p,n1 = 0,
A.nj-1 -(A.+ p,)nj + P,7rj+1 = 0 for j =:: 1,
with solutionni = no(A.jp,)i. We have in addition that:Li ni = 1 ifA < p,andno = {1- (A./p,)}- 1. 4.
One may use the strong Markov property. Alternatively, by the Markov property, lP'(Yn+1 = j I Yn = i, Tn = t, B)= lP'(Yn+1 = j I Yn = i, Tn = t)
for any event B defined in terms of {X (s) : s :S Tn}. Hence lP'(Yn+1 = j
I Yn =
i, B)= fooo lP'(Yn+1 = j
I Yn =
i, Tn = t)fTn(t)dt
= lP'(Yn+1 = j I Yn = i), so that Y is a Markov chain. Now qij = lP'(Yn+1 = j I Yn = i) is given by
% = { 00 Pij(t)A.e-M dt,
lo
294
Solutions
Continuous-time Markov chains
[6.9.5]-[6.9.8]
by conditioning on the (n + 1)th interarrival time of N; here, as usual, Pij (t) is a transition probability of X. Now
The jump chain Z = {Zn : n ::=: 0} has transition probabilities hij = 8i} / 8i, i -=f. j. The chance that Z ever reaches A from j is also IJj, and 11} = Lk hjkiJk for j as required.
5.
6. Let T1 = inf {t : X (t) -=f. X (0)}, and more generally let Tm be the time of the mth change in value of X. For j ¢:. A, /J-j = 1Ej(T1)
+ Lhjk/J-k> kf-j
where Ej denotes expectation conditional on Xo = j. Now Ej (Tl) = g1- 1 , and the given equations follow. Suppose next that (ak : k E S) is another non-negative solution of these equations. With Ui = Ti+1 - Ti and R = min{n 2: 1 : Zn E A}, we have for j
where
~
is a sum of non-negative terms. It follows that a}
2: Ej(Uo)
= Ej (
t
+ 1Ej(U1/{R>l)) + · · · + Ej(Unl(R>nJ) Url(R>r})
= Ej (min{Tn, HA})
-+ Ej(HA)
r=O
as n -+ oo, by monotone convergence. Therefore, f.L is the minimal non-negative solution.
7. First note that i is persistent if and only if it is also a persistent state in the jump chain Z. The integrand being positive, we can write
where {Tn : n ::=: 1} are the times of the jumps of X. The right side equals
1
00
00
L JE(T1 I X (0) = i)hu (n) = -:- L hu (n) n=O
gz n=O
where H = (hij) is the transition matrix of Z. The sum diverges if and only if i is persistent for Z.
8. Since the imbedded jump walk is persistent, so is X. The probability of visiting m during an excursion is a= (2m)- 1, since such a visit requires an initial step to the right, followed by a visit to m before 0, cf. Example (3.9.6). Having arrived at m, the chance of returning tom before visiting 0 is 1 -a, by the same argument with 0 and m interchanged. In this way one sees that the number N of visits tom during an excursion from 0 has distribution given by IP'(N ::=: k) = a(l - a)k- 1, k ::=: 1. The 'total jump rate' from any state is A., whence T may be expressed as :L~o Vi where the Vi are exponential with parameter A.. Therefore, JE(e eT )=GN ( -A.- ) ).-()
a). =(1-a)+a--. a).-()
295
[6.9.9]-[6.9.11]
Solutions
Markov chains
The distribution ofT is a mixture of an atom at 0 and the exponential distribution with parameter a)..
9. The number N of sojourns in i has a geometric distribution IP'(N = k) = fk-l(l- f), k =:: 1, for some f < 1. The length of each of these sojourns has the exponential distribution with some parameter 8i. By the independence of these lengths, the total time T in state i has moment generating function gi(l-f) 8i0- f ) -
e
The distribution ofT is exponential with parameter 8i (1 - f).
10. The jump chain is the simple random walk with probabilities ).j (). + JJ-) and JJ-/ (). + JJ-), and with POl = 1. By Corollary (5.3.6), the chance of ever hitting 0 having started at 1 is JJ-/A, whence the probability of returning to 0 having started there is f = JJ-/A. By the result of Exercise (6.9.9),
as required. Having started at 0, the walk visits the state r =:: 1 with probability 1. The probability of returning to r having started there is
and each sojourn is exponentially distributed with parameter 8r = ). + JJ-. Now g 7 (1 - fr) = ). - JJ-, whence, as above, JE(eevr) = ). - /JA-JJ--e The probability of ever reaching 0 from X (0) is (JJ-/A)x (O), and the time spent there subsequently is exponential with parameter ). - JJ-. Therefore, the mean total time spent at 0 is
11. (a) The imbedded chain has transition probabilities
where 8k = -gkk· Therefore, for any state j,
where we have used the fact thatJrG = 0. Also irk =:: 0 and L:k irk = 1, and thereforeJi is a stationary distribution of Y. Clearly fik = nk for all k if and only if 8k = l::i ni 8i for all k, which is to say that 8i = 8k for all pairs i, k. This requires that the 'holding times' have the same distribution. (b) Let Tn be the time of the nth change of value of X, with To= 0, and let Un = Tn+l - Tn. Fix a state k, and let H = min{n =:: 1 : Zn = k}. Let Yi (k) be the mean time spent in state i between two consecutive visits to k, and let Yi(k) be the mean number of visits to i by the jump chain in between two
296
Solutions
Birth-death processes and imbedding
[6.9.12]-[6.11.2]
visits to k (so that, in particular, Yk(k) = g/; 1 and n(k) = 1). With Ej and 1P'j denoting expectation and probability conditional on X (0) = j, we have that Yi(k) =lEk(f:unl(Zn=i,H>n}) = f:Ek(Un I f(Zn=iJ)lP'k(Zn = i, H > n) n=O n=O
1 1 -lP'k(Zn = i, H > n) = -yi(k). n=O 8i 8i 00
= L
The vector y(k) = (Yi (k) : i E S) satisfies y(k)H = y(k), by Lemma (6.4.5), where H is the transition matrix of the jump chain Z. That is to say, for j E S, whence I:i Yi (k)gij = 0 for all j E S. If tlk = I:i Yi (k) < oo, the vector (Yi (k)/ t-tk) is a stationary distribution for X, whence :rri = Yi(k)/t-tk for all i. Setting i = k we deduce that Trk = 1/(gktLk). Finally, ifi;i Tri8i < oo, then
~
1
tlk =-:::::--- = Trk
I:i Tri8i "' = tlk L:rri8i· Trk8k i
12. Define the generator G by gu = -vi, 8ij = vihij, so that the imbedded chain has transition matrix H. A root of the equation 1rG = 0 satisfies
0 = L:rri8ij = -TrjVj i
+
L (:rrivi)hij i:i-/-j
whence the vectors = (:rrjVj : j E S) satisfies s = tH. Therefore s = av, which is to say that Trj Vj = avj ,for some constant a. Now Vj > 0 for all j, so that Trj = a, which implies that I:j Trj =!= 1. Therefore the continuous-time chain X with generator G has no stationary distribution.
6.11 Solutions. Birth-death processes and imbedding 1.
The jump chain is a walk {Zn} on the setS= {0, 1, 2, ... } satisfying, fori ::: 1, lP'(Zn+I = j
I Zn = i) = {
if j = i + 1, ifj=i-1,
Pi 1- Pi
where Pi= Ai/(Ai + t-ti)· Also lP'(Zn+l = 1 I Zn = 0) = 1. 2.
The transition matrix H = (hij) of Z is given by i tL h .. - { A+it-t lJA -A+ it-t
'f .
.
1
1 ]=!-'
ifj=i+l.
To find the stationary distribution of Y, either solve the equation 1r = 1rQ, or look for a solution of the detailed balance equations :rrihi,i+l = :rri+Ihi+ 1,i. Following the latter route, we have that
297
[6.11.3]-[6.11.4]
Solutions
Markov chains
whence TCi =reo pi (1 + i j p )/ i! fori 2:: 1. Choosing reo accordingly, we obtain the result. It is a standard calculation that X has stationary distribution v given by vi = pie -p j i! fori 2:: 0. The difference between 1r and v arises from the fact that the holding-times of X have distributions which depend on the current state. 3.
We have, by conditioning on X(h), that
I
l](t +h)= lE{lP'(X(t +h) = 0 X(h))} = JJ-h · 1 + (1- A.h- JJ-h)l](t) + A.h~(t) + o(h) where~ (t) = lP'(X (t) = 0 I X (0) = 2). The process X may be thought of as a collection of particles each of which dies at rate fJ- and divides at rate A., different particles enjoying a certain independence; this is a consequence of the linearity of A.n and JJ-n. Hence ~ (t) = 11 (t ) 2 , since each of the initial pair is required to have no descendants at time t. Therefore
1J 1(t) = M- (A.+ JJ-)IJ(f) + A.17(t) 2
subject to 11 (0) = 0. Rewrite the equation as 11'
--,-,---------,___:_-----,-- = 1
(1 - IJ)(JJ- - A.IJ)
and solve using partial fractions to obtain if).=/)-,
Finally, if 0 < t < u, lP'(X(t) = 0 I X(u) = 0) = lP'(X(u) = 0 I X(t) = 0)
4.
lP'(X(t) = 0) lP'(X(u) = 0)
17(t) = -. 17(u)
The random variable X (t) has generating function JJ-(1 - s)- (M- A.s)e-t(A.-JL) G (s, t) = ;____:___ ___:____:__ ____:_..,.,-;------,).(1 - s)- (M- A.s)e-t(A.-JL)
as usual. The generating function of X(t), conditional on {X(t) > 0}, is therefore
~ sn lP'(X(t) = n) = G(s, t)- G(O, t). ~
lP'(X(t) > 0)
Substitute for G and take the limit as t
~
H(s)
=
1 - G(O, t)
oo to obtain as limit (M- A.)s /)--AS
oo
= Ls
n
Pn
n=l
where, with p = A./ JJ-, we have that Pn = pn-l (1 - p) for n 2:: 1.
298
Solutions
Special processes 5.
[6.11.5]-[6.12.1]
Extinction is certain if A. < JJ-, and in this case, by Theorem (6.11.10),
IfA > fJ- then lP'(T < oo) = JJ-/A., so JE(T
IT
< oo)
=
!oo
oo {
)..
1 - -lE(sX(t))l =O M s
}
dt
=
looo (A. - JJ-)e<~L-A)t ( -A)t dt o A. - JJ-e 11-
=
1 ( ).. ) -log - - . M A. - M
In the case A.= JJ-, lP'(T < oo) = 1 and JE(T) = oo.
6. By considering the imbedded random walk, we find that the probability of ever returning to 1 is max{A., JJ-}/(A. + JJ-), so that the number of visits is geometric with parameter min{A., JJ-}/(A. + JJ-). Each visit has an exponentially distributed duration with parameter A. + JJ-, and a short calculation using moment generating functions shows that V1 (oo) is exponential with parameter min{A., JJ-}. Next, by a change of variables, Theorem (6.11.10), and some calculation,
~srlE(Vr(t)) = lE(~
l
sr f(X(u)=r)du) = lE
(l
sX(u) du)
1 { A.(l- s)- (M- A.s)e-(A-~L)t} loot lE(s X( ) du = -JJ-fA. - -log A. A.-JJ1 { A.s(eP 1)} + terms not involving s, = --log 1 -
=
u )
1 -
1
A.
JJ-eP -A.
where p = JJ- - A.. We take the limit as t --+ oo and we pick out the coefficient of sr.
7.
If A.= JJ- then, by Theorem (6.11.10), lE(sX(t))
=
A.t(1 - s) A.t(l - s)
+s = 1 _ +1
1- s A.t(1 - s) + 1
and
r lE(sX(u)) du
lo
= t - _!_ log{A.t(l- s)
A.
+ 1}
1 1og { 1 - Ats = --} A. 1 +A.t
. . s. + terms not mvolvmg
Letting t--+ oo and picking out the coefficient of sr gives JE(Vr(oo)) = (rA.)- 1 . An alternative method utilizes the imbedded simple random walk and the exponentiality of the sojourn times.
6.12 Solutions. Special processes 1. The jump chain is simple random walk with step probabilities A./(A. expected time IJ-10 to pass from 1 to 0 satisfies
299
+ JJ-) and JJ-/(A. + JJ-).
The
[6.12.2]-[6.12.4]
Solutions
Markov chains
whence JJ,lQ = (JJ, + A.)/(JJ, -A.). Since each sojourn is exponentially distributed with parameter JJ, +A., the result follows by an easy calculation. See also Theorem (11.3.17). 2.
We apply the method of Theorem (6.12.11) with
the probability generating function of the population size at time u in a simple birth process. In the absence of disasters, the probability generating function of the ensuing population size at time v is
The individuals alive at timet arose subsequent to the most recent disaster at timet - D, where D has density function 8e- 8x, x > 0. Therefore,
3. The mean number of descendants after time t of a single progenitor at time 0 is e
lox
The aggregate effect at timex of N earlier arrivals is the same, by Theorem (6.12.7), as that of N arrivals at independent times which are uniformly distributed on [0, x]. Since JE(N) = vx, the mean population size at timex is v[e(A-~L)x- 1]/(A.- JJ,). The most recent disaster occurred at timet- D, where D has density function 8e - 8x, x > 0, and it follows that lE(X(t)) =
loo
t
v
8e- 8x--[e(A-11-)x-
A.-JJ,
1]dx
v 8 x[e(A-~L)t- 1]. + --e-
A.-JJ,
This is bounded as t --+ oo if and only if 8 > A. - JJ,.
4.
Let N be the number of clients who arrive during the interval [0, t]. Conditional on the event {N = n}, the arrival times have, by Theorem (6.12.7), the same joint distribution as n independent variables chosen uniformly from [0, t]. The probability that an arrival at a uniform time in [0, t] is still in service at timet is f3 = f~[l - G(t- x)]t- 1 dx, whence, conditional on {N = n}, the total number M still in service is bin(n, {J). Therefore,
whence M has the Poisson distribution with parameter A.{Jt = A. f~[l - G(x)] dx. Note that this parameter approaches A.JE(S) as t --+ oo.
300
Solutions
Spatial Poisson processes
[6.13.1]-[6.13.3]
6.13 Solutions. Spatial Poisson processes 1. It is easy to check from the axioms that the combined process N(t) = B(t) + G(t) is a Poisson process with intensity f3 + y. (a) The time S (respectively, T) until the arrival of the first brown (respectively, grizzly) bear is exponentially distributed with parameter f3 (respectively, y ), and these times are independent. Now, lP'(S < T) = {oo f3e-{Jse-ys ds = _f3__
lo
f3 + Y
(b) Using (a), and the lack-of-memory of the process, the required probability is
(c) Using Theorem (6.12.7),
I
lE(min{S, T} B(l)
1 } G(1)+2
= 1) = lE { .
=
y- 1 + e-Y y2
2. Let B, be the ball with centre 0 and radius r, and let Nr = ITI (6.13.11) that S, = Z::xEnnB, g(x) satisfies
JE(S, IN,)= N, where A(B) = fyEB A.(y) dy. Therefore, JE(S,) gence that JE(S) = JJRd g(u)A.(u) du. Similarly,
JE(s;
1
N,) = lE
(l L
1 Br
=
n Brl-
·
We have by Theorem
A.(u) g(u)--du, A(B,)
fsr g(u)A.(u) du, implying by monotone conver-
2
g(x)J )
EnnBr
~ c~H, E
g(x)
2))
+E (
~
g(x)g(y))
x,yEnnBr = N,
1
A.(u) g(u) - - du
Br
whence lE(S;) =
2
A(Br)
+ N, (Nr
- 1)
11u
g(u)g(v)
u,vEBr
A.(u)A.(v) A(Br)
2 du dv,
r
fsr g(u) A.(u) du + (fsr g(u)A.(u) du 2
By monotone convergence,
and the formula for the variance follows. 3. If B 1 , B2, ... , Bn are disjoint regions of the disc, then the numbers of projected points therein are Poisson-distributed and independent, since they originate from disjoint regions of the sphere. By
301
[6.13.4]-[6.13.8]
Solutions
Markov chains
elementary coordinate geometry, the intensity function in plane polar coordinates is 2'A/ ~. < 2n.
0 ::::; r ::::; 1, 0 ::::; 4.
e
The same argument is valid with resulting intensity function 2'A ~-
The Mercator projection represents the spherical coordinates (e, ¢)as Cartesian coordinates in the range 0 ::::; ¢ < 2n, 0 ::::; n . (Recall that is the angle made with the axis through the north pole.) Therefore a uniform intensity on the globe corresponds to an intensity .function 'A sine on the map. Likewise, a uniform intensity on the map corresponds to an intensity 'A/ sine on the globe.
5.
e ::;
e
6.
Let the Xr have characteristic function¢. Conditional on the value of N(t), the corresponding arrival times have the same distribution as N(t) independent variables with the uniform distribution, whence JE(eieS(t)) = lE{lE(eieS(t) I N(t))} = lE{lE(eiexe-au )N(t)}
= exp{A.t(JE(ewxe-au)- I)}= exp{ 'A
lot {¢(ee-"u)- 1} du }·
where U is uniformly distributed on [0, t]. By differentiation,
Now, for s < t, S(t) = S(s )e-a(t-s) + S(t- s) where S(t- s) is independent of S(s) with the same distribution as S(t- s). Hence, for s < t, 'AlE(X 2 ) 'AJE(X 2 ) cov(S(s), S(t)) = ~a(t-s) = ~(1- e-2as)e-a(t-s)--+ ~e-av ass --+ oo with v = t - s fixed. Therefore, p(S(s), S(s + v))--+ e-av ass--+ oo.
7.
The first two arrival times T1 , Tz satisfy
Differentiate with respect to x andy to obtain the joint density function A.(x)'A(x
+ y)e-A(x+y),
x, y ::: 0. Since this does not generally factorize as the product of a function of x and a function of y, T1 and Tz are dependent in general. 8.
Let Xi be the time of the first arrival in the process Ni. Then lP'(I
=I, T:::
t)
= lP'(t::::; x 1
=
1
00
< inf{X 2 ,
x3 , ... })
lP'(inf{Xz, X3, ... } > x)'A1e-Jqx dx
t
302
=
Al e-J..t_ 'A
Solutions
Markov chain Monte Carlo
[6.14.1]-[6.14.4]
6.14 Solutions. Markov chain Monte Carlo 1.
If P is reversible then
RHS
= L(~PijXj )YiJl"i = ~JriPijXjYi = ~JrjPjiYiXj l
J
l,]
l,]
=
~njXj (~PjiYi) J
= LHS.
l
Suppose conversely that (x, Py) = (Px, y) for all x, y E l 2 (n). Choose x, y to be unit vectors with 1 in the ith and jth place respectively, to obtain the detailed balance equations ni Pij = nj Pji. 2. Just check that 0 ::::; bij ::::; 1 and that the Pij = 8ij bij satisfy the detailed balance equations (6.14.3).
3. It is immediate that Pjk = IAjkl, the Lebesgue measure of Ajk· This is a method for simulating a Markov chain with a given transition matrix. 4. (a) Note first from equation (4.12.7) that d(U) = ~ supi#j dTv(ui-o Uj.), where Ui. is the mass function Uif, t E T. The required inequality may be hacked out, but instead we will use the maximal coupling of Exercises (4.12.4, 5); see also Problem (7.11.16). Thus requires a little notation. For i, j E S, i =f. j, we find a pair (Xi, Xj) of random variables taking values in T according to the marginal mass functions Ui., Uj., and such that IP'(Xi =f. Xj) = ~dTv(uh Uj.). The existence of such a pair was proved in Exercise (4.12.5). Note that the value of Xi depends on j, but this fact has been suppressed from the notation for ease of reading. Having found (Xi, Xj ), we find a pair (Y(Xi), Y(Xj)) taking values in U according to the marginal mass functions vx;·• vxj'' and such that IP'(Y(Xi)
=f.
Y(Xj)
I Xi,
Xj) =
IP'(Y(Xi)
=f.
~dTV(vx;., vxr). Now, taking a further liberty with the notation,
Y(Xj)~=
IP'(Xi = r, Xj = s)IP'(Y(r)
L
=f. Y(s))
,sES
is = L IP'(Xi = r, Xj = s)idTV(Vr., Vs.) r,sES ri=s
::'0
g supdTV(Vr., Vs.)}IP'(Xi =f. Xj), ri=s
whence d(UV) = supiP'(Y(Xi)
=f.
Y(Xj)) ::'0
ii=j
g rfs supdTV(Vr., Vs.) }{sup!P'(Xi =f. Xj)} i,j
and the claim follows. (b) WriteS= {1, 2, ... , m}, and take
u=
(IP'(Xo = 1) IP'(Yo = 1)
IP'(Xo = 2) IP'(Yo = 2)
m))
IP'(Xo = IP'(Yo = m)
·
The claim follows by repeated application of the result of part (a). It may be shown with the aid of a little matrix theory that the second largest eigenvalue of a finite stochastic matrix Pis no larger in modulus that d(P); cf. the equation prior to Theorem (6.14.9).
303
[6.15.1]-[6.15.6]
Solutions
Marlwv chains
6.15 Solutions to problems 1. (a) The state 4 is absorbing. The state 3 communicates with 4, and is therefore transient. The set {1, 2} is finite, closed, and aperiodic, and hence ergodic. We have that /34(n) = Ct)n- 1 ~'so that /34 = l:n /34(n) = ~(b) The chain is irreducible with period 2. All states are non-null persistent. Solve the equation 1f = 1f P to find the stationary distribution 1f = ( ~, ft, 156 , ~) whence the mean recurrence times are 8 16 16 8 . d 3 , 3 , 3"• , m or er.
2.
(a) Let P be the transition matrix, assumed to be doubly stochastic. Then LPij(n) i
= LLPik(n -OPkj = "L(LPik(n -l))Pkj i
k
k
i
whence, by induction, the n-step transition matrix pn is doubly stochastic for all n :::: 1. If j is not non-null persistent, then Pij (n) --+ 0 as n --+ oo, for all i, implying that l::i Pij (n) --+ 0, a contradiction. Therefore all states are non-null persistent. If in addition the chain is irreducible and aperiodic then Pij (n) --+ lij, where 1f is the unique stationary distribution. However, it is easy to check that 1f = (N- 1 , N- 1 , ... , N- 1) is a stationary distribution if Pis doubly stochastic. (b) Suppose the chain is persistent. In this case there exists a positive root of the equation x = xP, this root being unique up to a multiplicative constant (see Theorem (6.4.6) and the forthcoming Problem vector of 1's. By the (7)). Since the transition matrix is doubly stochastic, we may take x = 1, ~e above uniqueness of x, there can exist no stationary distribution, and there£ e the chain is null. We deduce that the chain cannot be non-null persistent. 3.
By the Chapman-Kolmogorov equations, m,r,n:::: 0.
Choose two states i and j, and pick m and n such that a= Pij (m)Pji(n) > 0. Then Pu(m
+ r + n):::: apjj(r).
Set r = 0 to find that Pii (m + n) > 0, and so d(i) I (m + n). If d(i) f r then Pii (m + r that Pjj(r) = 0; therefore d(i) I d(j). Similarly d(j) I d(i), giving that d(i) = d(j).
+ n) =
0, so
4. (a) See the solution to Exercise (6.3.9a). (b) Let i, j, r, s E S, and choose N(i, r) and N(j, s) according to part (a). Then
I
lP'(Zn = (r, s) Zo = (i,
n)
= Pir(n)Pjs(n) > 0
if n:::: max{N(i, r), N(j, s)}, so that the chain is irreducible and aperiodic. (c) SupposeS= {1, 2} and
P=C 6)· In this case { {1, 1}, {2, 2}} and { {1, 2}, {2, 1}} are closed sets of states for the bivariate chain. 5. Clearly lP'(N = 0) = 1 - /ij, while, by conditioning on the time of the nth visit to j, we have that lP'(N :::: n + 1 I N :::: n) = fjj for n :::: 1, whence the answer is immediate. Now lP'(N = oo) = 1 - l:~o lP'(N = n) which equals 1 if and only if /ij = fjj = 1. 6. Fix i =/= j and let m = min{n : Pij (n) > 0}. If Xo = i and Xm = j then there can be no intermediate visit to i (with probability one), since such a visit would contradict the minimality of m.
304
Solutions
Problems
[6.15.7]-[6.15.8]
Suppose Xo = i, and note that (1 - f}i) Pij (m) :5 1 - Iii, since if the chain visits j at time m and subsequently does not return to i, then no return to i takes place at all. However fii = 1 if i is persistent, so that fj i = 1. 7.
(a) We may takeS= {0, 1, 2, ... }. Note that qij (n)::: 0, and 00
L%(n) =
%(n
1,
+ 1) = Lqu(l)qu(n),
j
1=0
whence Q = (% (1)) is the transition matrix of a Markov chain, and Qn = (% (n)). This chain is persistent since for all i, n
n
and irreducible since i communicates with j in the new chain whenever j communicates with i in the original chain. That i =I= j, n ::: 1,
is evident when n = 1 since both sides are qij (1). Suppose it is true for n = m where m ::: 1. Now lji(m
L
+ 1) =
i =I= j,
ljk(m)Pki·
k:kfj
so that
x·
_l_ lji(m
L
+ 1) =
Xi
i =I= j,
gkj(m)qik(1),
k:kfj
which equals gij (m + 1) as required. (b) Sum (*) over n to obtain that i =I= j,
where Pi(j) is the mean number of visits to i between two visits to j; we have used the fact that I:n gij (n) = 1, since the chain is persistent (see Problem (6.15.6)). It follows that Xi = XOPi (0) for all i, and therefore x is unique up to a multiplicative constant. (c) The claim is trivial when i = j, and we assume therefore that i =I= j. Let Ni (j) be the number of visits to i before reaching j for the first time, and write lP'k and lEk for probability and expectation conditional on Xo = k. Clearly, lP'j(Ni(j):::: r) = hjiO- hijy-l for r::: 1, whence
The claim follows by(**).
8.
(a) If such a Markov chain exists, then n
Un =
L
fiUn-i•
i=l
305
n:::
1,
[6.15.9]-[6.15.9] where
Solutions
Marlwv chains
fi is the probability that the first return of X to its persistent starting point s takes place at time
i. Certainly uo = 1. Conversely, suppose u is a renewal sequence with respect to the collection Um : m 2: 1). Let X be a Markov chain on the state space S = {0, 1, 2, ... } with transition matrix
.. _ { lP'(T 2: i + 2 I T 2: i + 1) PI] 1 - lP'(T 2: i + 2 I T 2: i + 1)
if j = i + 1, if j = 0,
where Tis a random variable having mass function fm = lP'(T = m). With Xo = 0, the chance that the first return to 0 takes place at time n is
lP'( Xn = 0,
IT
xi
i= 0 I Xo
= 0) = P01P12 ... Pn-2,n-1Pn-1,0
1
IT
= ( 1 _ G(n + 1)) G(i + 1) G(n) i=l G(i)
= G(n) -
G(n
+ 1) =
fn
where G(m) = lP'(T 2: m) = L:~m fn· Now Vn = lP'(Xn = 0 I Xo = 0) satisfies n
vo
= 1,
forn::;::1,
Vn = LfiVn-i i=l
whence Vn = Un for all n. (b) Let X andY be the two Markov chains which are associated (respectively) with u and v in the above sense. We shall assume that X andY are independent. The product (unvn: n 2: 1) is now the renewal sequence associated with the bivariate Markov chain (Xn, Yn).
2n (2n)
9. Of the first steps, let there be i rightwards, j upwards, and k inwards. Now Xzn = 0 if and only if there are also i leftwards, j downwards, and k outwards. The number of such possible !/ {(i! j! k!) 2 }, and each such combination has probability (t) 2U+ j+k) = (t) 2 combinations is The first equality follows, and the second is immediate. Now
n.
1 ) 2n lP'(Xzn = 0) ::S ( 2 where
M=max{ 3 n·~!·lkl: l. J . .
(2n) M n
L
i+J+k=n
n! 3n·1 "lkl
l.J. ·
i,j,k::;::O, i+j+k=n}.
It is not difficult to see that the maximum M is attained when i, j, and k are all closest to :} n, so that
Furthermore the summation in (*) equals 1, since the summand is the probability that, in allocating n balls randomly to three urns, the urns contain respectively i, j, and k balls. It follows tliat (2n)! lP'(X2 = 0) < -----'-----7----::n
- 12nn!
306
(Lj-nJ !)3
Solutions [6.15.10]-[6.15.13]
Problems
3
which, by an application of Stirling's formula, is no bigger than Cn -2 for some constant C. Hence l:n lP'(X2n = 0) < oo, so that the origin is transient.
10. No. The line of ancestors of any cell-state is a random walk in three dimensions. The difference between two such lines of ancestors is also a type of random walk, which in three dimensions is transient.
11. There are one or two absorbing states according as whether one or both of a and f3 equal zero. If af3 =f. 0, the chain is irreducible and persistent. It is periodic if and only if a = f3 = 1, in which case it has period 2.
IfO < af3 < 1 then 1r=
(a!f3'a:f3)
is the stationary distribution. There are various ways of calculating pn; see Exercise (6.3.3) for example. In this case the answer is given by
proof by induction. Hence as n ---+ oo. The chain is reversible in equilibrium if and only if n1 P12
= 7r2P21, which is to say that af3 = f3a
!
12. The transition matrix is given by
Pi}=
(NN-i)2 ifj=i+1, (NN- i) 2 if j = i, 1 - (Ni ) if j = i - 1, (Ni )2 2
_
for 0 :::; i :::; N. This process is a birth-death process in discrete time, and by Exercise (6.5.1) is reversible in equilibrium. Its stationary distribution satisfies the detailed balance equation Jri Pi,i+l = ni+lPi+l,i for 0:::; i < N, whence ni = no(~) 2 for 0:::; i :::; N, where
_!_
no
=
t (~) i=O
z
2
=
(2N). N
13. (a) The chain X is irreducible; all states are therefore of the same type. The state 0 is aperiodic, and so therefore is every other state. Suppose that Xo = 0, and let T be the time of the first return to 0. Then lP'(T > n) = aoa1 · · · an-1 = bn for n :::: 1, so that 0 is persistent if and only if bn ---+ 0 as n ---+ oo. (b) The mean of T is 00
JE(T)
=
00
L lP'(T > n) = L bn. 307
[6.15.14]-[6.15.14]
Solutions
Markov chains
The stationary distribution 1r satisfies 00
no= L 7l'k(l- ak), k=O
Jl'n = 7l'n-1Gn-1 for n 2: 1.
Hence Jl'n = nobn and n 1 = :L~o bn if this sum is finite. (c) Suppose ai has the stated form for i 2: I. Then
0
n-1
II (1 -
bn =hi
Ai-f3),
n 2: /.
i=I
Hence bn --+ 0 if and only if :Li Ai-fJ = oo, which is to say that persistent if and only if f3 ::=: 1. (d) We have that 1 - x :::: e-x for x 2: 0, and therefore
f3 ::=:
1. The chain is therefore
if{J
(e) If f3 = 1 and A > 1, there is a constant c I such that oo oo { n-1 1 } oo oo Lbn::SbiLexp - A L i ::SciLexp{-Alogn}=ciLn-A
giving that the chain is non-null. (f) If f3 = 1 and A :::: 1,
bn = b I
n-1( A) II 1 - --;-
i=I
2: b I
n-1(. II ~1) i=I
l
= bI
l
(/ =1) . n
1
Therefore :Ln bn = oo, and the chain is null.
14. Using the Chapman-Kolmogorov equations,
IPi} (t + h) -
Pij (t) I =
IL (Pik (h) -
Oik) Pkj (t)
I :::: (1 -
Pii (h)) Pij (t)
::S
(1 -
+ L Pik (h) k=Ji
k Pii (h))
+ (1 -
Pii (h)) --+ 0
as h .j, 0, if the semigroup is standard. Now logx is continuous for 0 < x :::: 1, and therefore g is continuous. Certainly g(O) = 0. In addition Pii (s + t) 2: Pii (s)pii (t) for s, t 2: 0, whence g(s + t) ::S g(s) + g(t), s, t 2: 0. For the last part
1 ((Pii(t) - 1)
=
g(t) Pii (t) - 1 -t- · -log{1- (1- pu(t))}--+ -A.
as t .j, 0, since x flog( 1 - x) --+ -1 as x .j, 0.
308
Solutions
Problems
[6.15.15]-[6.15.16]
15. Let i and j be distinct states, and suppose that Pij (t) > 0 for some t. Now
implying that (Gn)ij > 0 for some n, which is to say that
for some sequence k1, k2, ... , kn of states. Suppose conversely that (*)holds for some sequence of states, and choose the minimal such value of n. Then i, k1, k2, ... , kn, j are distinct states, since otherwise n is not minimal. It follows that (Gn)ij > 0, while (Gm)ij = 0 for 0::: m < n. Therefore 00
Pij (t) = t
n"'l
k
k
L..J k! t (G )ij
k=n
is strictly positive for all sufficiently small positive values oft. Therefore i communicates with j.
16. (a) Suppose X is reversible, and let i and j be distinct states. Now lP'(X(O) = i, X(t) =
j)
= lP'(X(t) = i, X(O) =
j),
which is to say that Tri Pij (t) = Trj Pji (t). Divide by t and take the limit as t
+0 to obtain that
Trigij = Trjgji·
Suppose now that the chain is uniform, and X (0) has distribution 1r. If t > 0, then lP'(X (t) = j) =
2.:: Tri Pij (t) = Trj, i
so that X(t) has distribution 1r also. Now lett < 0, and suppose that X(t) has distribution J.L. The distribution of X (s) for s ::: 0 is J.LPs-t = 1r, a polynomial identity in the variables - t, valid for all s ::: 0. Such an identity must be valid for all s, and particularly for s = t, implying that J.L = 1r. Suppose in addition that Tri gij = Trj gji for all i, j. For any sequence k1, k2, ... , kn of states, Trigi,ktgkt,kz ···gkn,j = gkJ,iTrktgkt,kz · · ·gkn,j = ··· = gk1,igkz,k1 ·· ·gj,knTrj.
Sum this expression over all sequences k1, k2, ... , kn oflength n, to obtain Tri(Gn+1)ij = Trj(Gn+1)ji•
n::: 0.
It follows, by the fact that Pt = etG, that
foralli,j,t. Fort1 < t2 < ··· < tn, lP'(X(t1) = i1, X(t2) = i2, ... , X(tn) =in)
= Tri! Pit ,iz (t2 - t1) · · · Pin_ 1,in (tn - tn-1) = Piz,i1 (t2- tl)TrizPiz,i3 (t3- t2) · · · Pin-J,in Ctn- tn-1) = · · · = Pi2 ,i 1 (t2- t1) · · · Pin.in- 1 Ctn- tn-1)1rin = lP'(Y(t1) = i1, Y(t2) = i2, ... , Y(tn) =in),
309
[6.15.17]-[6.15.19]
Markov chains
Solutions
giving that the chain is reversible. (b) LetS = {1, 2} and
G
=(-a a) f3
-{3
where af3 > 0. The chain is uniform with stationary distribution
and therefore Jqgl2 = n2g2i· (c) Let X be a birth--death process with birth rates A.; and death rates J.Li. The stationary distribution n satisfies
Therefore nk+iiLk+i = Jl'kAk fork 2: 0, the conditions for reversibility.
17. Consider the continuous-time chain with generator G = (-{3
y
f3 ) .
-y
It is a standard calculation (Exercise (6.9.1)) that the associated semigroup satisfies ({3
+
+ f3h(t) y(1- h(t))
)P = ( y y
t
{3(1- h(t))) f3 + yh(t)
where h(t) = e-t(,B+y). Now P1 = Pifandonlyify + f3h(1) = f3 + yh(1) to say that f3 = y = -1log(2a - 1), a solution which requires that a > 1·
= a(f3 +
y), which is
18. The forward equations for Pn(t) = lP'(X(t) = n) are Pb(t) = JLPi - A.po,
P~ (t) = APn-i - (A.+ nJL) Pn + JL(n + 1) Pn+i, In the usual way,
aa at
=(s- 1)(A.G- 11
no::
I.
aa) as
with boundary condition G(s, 0) =sf. The characteristics are given by
ds
dt= - - JL(S - 1)
dG A.(s - 1)G'
and therefore G = eP(s-i) f ( (s-l)e-JU), for some function f, determined by the boundary condition to satisfy eP(s-l) f(s- 1) =sf. The claim follows. As t-+ oo, G(s, t)-+ eP(s-i), the generating function of the Poisson distribution, parameter p.
19. (a) The forward equations are
a
-p;;(s, t) = -A.(t)p;;(s, t), at
a
atPij(s,t) = -A.(t)pij(S, t) +A.(t)Pi,j-J(t),
310
i < j.
Solutions [6.15.20]-[6.15.20]
Problems Assume N(s) = i and s < t. In the usual way, 00
G(s, t; x)
= 2:xilP'(N(t) = j
I N(s)
= i)
J=i
satisfies iJG
at= A.(t)(x- 1)G.
We integrate subject to the boundary condition to obtain G(s, t; x) =xi exp { (x- 1)
1t
A.(u) du},
whence Pi} (t) is found to be the probability that A = j - i where A has the Poisson distribution with
J:
parameter A.(u) du. The backward equations are
a
osPij(S, t) = A.(s)Pi+l,j(S, t)- A.(s)Pij(S, t);
using the fact that Pi+ I,} (t) = Pi,J-1 (t), we are led to iJG - - = A.(s)(x- !)G.
OS
The solution is the same as above. (b) We have that lP'(T > t)
= Poo(t) = exp {
so that fT(t) = A.(t)exp {
In the case A.(t) = cj(l
-l
-
lot
A.(u) du},
t 2:0.
A.(u)du }•
+ t), E(T)
= !o oo lP'(T 0
> t)dt
= !ooo o
du
(1
+ u)c
which is finite if and only if c > 1. 20. Let s > 0. Each offer has probability 1 - F (s) of exceeding s, and therefore the first offer exceeding sis the Mth offer overall, where lP'(M = m) = F(s)m-l [I - F(s)], m 2: 1. Conditional on {M = m}, the value of XM is independent of the values of X1, X2, ... , XM-1, with lP'(XM > u I M = m) =
1- F(u) , I - F(s)
0<
S ::": U,
and X 1, X 2, ... , X M -I have shared (conditional) distribution function F(u) G(u Is)= F(s),
311
0:::: u:::: s.
[6.15.21]-[6.15.21]
Solutions
Markov chains
For any event B defined in terms of X 1, X 2, ... , X M -! , we have that 00
IP'(XM > u, B)=
L
IP'(XM > u, B
IM =
m)IP'(M = m)
m=i 00
=
L
IM =
IP'(XM > u
m)IP'(B
IM =
m)IP'(M = m)
m=i 00
= IP'(XM > u)
L IP'(B I M = m)IP'(M = m) m=i
0<
= IP'(XM > u)IP'(B),
S _:::: U,
where we have used the fact that IP'(XM > u I M = m) is independent of m. It follows that the first record value exceeding s is independent of all record values not exceeding s. By a similar argument (or an iteration of the above) all record values exceeding s are independent of all record values not exceeding s. The chance of a record value in (s, s + h] is IP'(s < XM < s +h)= -
F(s +h)- F(s)
1- F(s)
=
J(s)h
1- F(s)
+o(h).
A very similar argument works for the runners-up. Let XM 1 , XM2 , .•. be the values, in order, of offers exceeding s. It may be seen that this sequence is independent of the sequence of offers not exceeding s, whence it follows that the sequence of runners-up is a non-homogeneous Poisson process. There is a runner-up in (s, s + h] if (neglecting terms of order o(h)) the first offer exceeding s is larger than s + h, and the second is in (s, s + h]. The probability of this is ( 1-F(s+h)) (F(s+h)-F(s)) +o(h ) = f(s)h +o (h) . 1-F(s) 1-F(s) 1-F(s)
21. Let F 1 (x) = IP'(N*(t) _:::: x), and let A be the event that N has a arrival during (t, t +h). Then Fr+h(x) = A.h!P'(N*(t +h)_::::
xI A)+ (1- A.h)Fr(x) + o(h)
where IP'(N*(t +h)_::::
Hence
~Fr(x) = at
xI A)=
-A.Fr(x)
+
L:
t.../
00
-oo
Fr(x- y)f(y)dy.
Fr(x- y)f(y)dy.
Take Fourier transforms to find that ¢ 1 (8) = IE(eifiN*(t)) satisfies
an equation which may be solved subject to ¢o(8) = 1 to obtain ¢ 1 (8) = eAI(>(e)-ll. Alternatively, using conditional expectation,
where N (t) is Poisson with parameter A.t.
312
Solutions [6.15.22]-[6.15.25]
Problems
22. We have that E(sN(t)) = E{E(sN(t)
I A)}=
1{eJ.It(s-1)
+ eJ.2t(s-l)},
= ~(A.J + A.z)t and var(N(t)) = 1(A.J + A.z)t + i
whence E(N(t)) 23.
24. The forward equations for Pn(t) = li"(X(t) = n) are
n 2: 0, with the convention that P-i (t) = 0. Multiply by sn and sum to deduce that (1
aa
aa
aG
2 + J.Lf)at = sG + JLS as - G- JLSas
as required. Differentiate with respect to s and take the limit ass m(t)
t
l. If E(X (t) 2 ) < oo, then
= E(X(t)) = -aal
as
s=i
satisfies (1 + J.Lt)m 1(t) = 1 + J.Lm(t) subject to m(O) =I. Solving this in the usual way, we obtain m(t) = I+ (1 + JL/)t. Differentiate again to find that n(t) = E(X(t)(X(t)
satisfies (1
+ J.Lf)n 1 (t)
= 2(m(t)
+ J.Lm(t) + J.Ln(t))
n(t) = I (I - 1)
The variance of X(t) is n(t)
-1))
a2al
= -
as
2
s=i
subject to n(O) =I (I- 1). The solution is
+ 2I (1 + JL/)t + (1 + 11 I)(l + 11 + 11 I)t 2 .
+ m(t) -
m(t) 2 .
25. (a) Condition on the value of the first step: TJ} =
AJ
"}
as required. Set Xi
= TJi+i
ILJ
..,--------+ . TJ} +I + ..,--------+ . TJ} -!, !Lj
"}
- TJi to obtain AJXJ j
XJ
!Lj
= JLjXJ-i
for j 2: 1, so that
w
=xoiT f.· i=i l
It follows that TJ}+i
= TJO +
j
j
k=O
k=O
L Xk = 1 + (TJJ - 1) L ek.
The TJ} are probabilities, and lie in [0, 1]. If l::f ek = oo then we must have TJJ = 1, which implies that TJ} = 1 for all j.
313
[6.15.26]-[6.15.28]
Solutions
Markov chains
(b) By conditioning on the first step, the probability T/J, of visiting 0 having started from j, satisfies
u + I) 2 TJj+i + p,j-1 j2
T/j =
+ (j + 1)2
Hence, (j + 1) 2 (TJj+i- TJj) = P(TJj- T/j-J), giving (j + 1) 2 (TJj+i- TJj) = TJi- TJO· Therefore, j
1- TJj+i = (1- TJJ) '"" L.J
1
k=O (k
+ 1)
I 2 2 -+ (1- TJI)6n
as j-+ oo.
By Exercise (6.3.6), we seek the minimal non-negative solution, which is achieved when T/1 = 1 (6/n 2).
26. We may suppose that X(O) = 0. Let Tn = inf{t : X(t) = n}. Suppose Tn = T, and let Y = Tn+i - T. Condition on all possible occurrences during the interval (T, T +h) to find that E(Y) = (A.nh)h + f.lnh(h + E(Y 1)) + (1 - A.nh- P,nh)(h + E(Y)) + o(h), where Y' is the mean time which elapses before reaching n + 1 from n- 1. Set mn = E(Tn+i - Tn) to obtain that mn = P,nh(mn-i + mn) + mn + h{l- (A.n + P,n)mn} + o(h). Divide by h and take the limit ash
.,!..
0 to find that A.nmn = 1 + Jlnmn-i, n 2: 1. Therefore
1 f.ln 1 Jln Jlnf.ln-i ···Ill mn = - + -mn-i = · · · = - + - - - + · · · + • An An An An An-i An An-i · · · A.o
since mo = A. 01 . The process is dishonest if ~~ 0 mn < oo, since in this case T00 = lim Tn has finite mean, so that lP'(T00 < oo) = 1. On the other hand, the process grows no faster than a birth process with birth rates Ai, which is honest if ~~ 0 1/A.n = oo. Can you find a better condition? 27. We know that, conditional on X (0) = I, X (t) has generating function
A.t(l-s)+s)I G(s, t) = ( A.t(l - s) + 1 , so that lP'(T:::; x
1
A.x X(O) =I)= lP'(X(x) = 0 I X(O) =I)= G(O, x) = ( A.x + 1
)I
It follows that, in the limit as x -+ oo, lP'(T :S x) = ~ L.J ( -A.xI=O Ax + 1
)I
lP'(X(O) = /) = Gx(O)
(
-A.x- ) -+ 1. Ax + 1
For the final part, the required probability is {xI I (xI + 1)} I = { 1 + (x /)-I }-I, which tends to e-ijx as I -+ oo.
28. Let Y be an immigration-death process without disasters, with Y (0) = 0. We have from Problem (6.15.18) that Y(t) has generating function G(s, t) exp{p(s- 1)(1- e-!Lt)} where p A.fp,. As seen earlier, and as easily verified by taking the limit as t -+ oo, Y has a stationary distribution.
=
314
=
Solutions [6.15.29]-[6.15.31]
Problems
From the process Y we may generate the process X in the following way. At the epoch of each disaster, we paint every member of the population grey. At any given time, the unpainted individuals constitute X, and the aggregate population constitutes Y. When constructed in this way, it is the case that Y(t) :::; X(t), so that Y is a Markov chain which is dominated by a chain having a stationary distribution. It follows that X has a stationary distribution :n: (the state 0 is persistent for X, and therefore persistent for Y also). Suppose X is in equilibrium. The times of disasters form a Poisson process with intensity 8. At any given time t, the elapsed time T since the last disaster is exponentially distributed with parameter 8. At the time of this disaster, the value of X (t) is reduced to 0 whatever its previous value. It follows by averaging over the value ofT that the generating function H(s) = L~o snnn of X (t) is given by
by the substitution x = e-JLu. The mean of X (t) is
29. Let G(IBI, s) be the generating function of X(B). If BnC = 0, then X(B UC) = X(B)+X(C), so that G(a + f.l, s) = G(a, s)G(f.l, s) for Is I :::; 1, a, f.l ::=: 0. The only solutions to this equation which are monotone in a are of the form G(a, s) = eaJ..(s) for lsi :::; 1, and for some function A.(s). Now any interval may be divided into n equal sub-intervals, and therefore G(a, s) is the generating function of an infinitely divisible distribution. Using the result of Problem (5.12.13b), A.(s) may be written in the form A.(s) = (A(s) -l)A. for some A. and some probability generating function A(s) = L~ aisi. We now use (iii): if lEI= a, IP(X (B) :0:: 1) IP(X(B) = 1) as a .,), 0. Therefore ao + a 1 = 1, and hence A(s) = ao distribution with parameter proportional to IB 1.
+ (1- ao)s,
and X(B) has a Poisson
30. (a) Let M (r, s) be the number of points of the resulting process on R+ lying in the interval (r, s]. Since disjoint intervals correspond to disjoint annuli of the plane, the process M has independent increments in the sense that M(rJ, SJ), M(r2, s2), ... , M(rn, sn) are independent whenever r1 < SJ < r2 < · · · < rn < Sn. Furthermore, for r
r)}ke-J..n(s-r)
k!
.
(b) We have similarly that
IP(R(k):::; x) = IP(N has leastk points in circle of radius x) =
Loo
(A.nx2t e-J..nx2
r= k
r.1
,
and the claim follows by differentiating, and utilizing the successive cancellation. 31. The number X(S) of points within the sphere with volume Sand centre at the origin has the Poisson distribution with parameter A.S. Hence IP(X (S) = 0) = e-J..S, implying that the volume V of the largest such empty ball has the exponential distribution with parameter A..
315
[6.15.32]-[6.15.33]
Solutions
Markov chains
It follows that lP'(R > r) = JP'(V > ern) = ball in n dimensions. Therefore
e-J...crn
for r 2: 0, where c is the volume of the unit
r 2: 0.
Finally, E(R) = J000 e-J...crn dr, and we set v = 'Acrn.
32. The time between the kth and (k + 1)th infection has mean 'Ak" 1, whence N
E(T)
1
=I:-. Ak k=1
Now N
1
{N
1
N
1
.(.; k(N + 1 - k) = N + 1 .(.;
k + .(.; N
1 } + 1- k
1 2 L= --{logN +y +O(N- 1)}. N +1 k N +1 2
= --
N
k= 1
It may be shown with more work (as in the solution to Problem (5.12.34)) that the moment generating function of 'A(N + l)T- 2log N converges as N-+ oo, the limit being {r(l- 8)} 2 .
33. (a) The forward equations for Pn (t) = JP'(V (t) = n + ~) are p~(t)
= (n + 1)Pn+l(t)- (2n + l)pn(t) +nPn-1(t),
n 2: 0,
with the convention that P-1 (t) = 0. It follows as usual that a a= aaat
( 2s-+G aa )
as
as
+
( s 2 -+sG aa ) as
as required. The general solution is G(s,
t) =
- 1-J
1-s
1- ) (t + -1-s
for some function f. The boundary condition is G(s, 0) = 1, and the solution is as given. (b) Clearly mn(T)=E(foT lntdt) =loT ECint)dt
by Fubini's theorem, where lnt is the indicator function of the event that V (t) = n + ~. As for the second part,
Loo snmn(T) = n=O
r G(s, lo T
t) dt = log[ 1 + (1 - s)T], 1- s
so that, in the limit as T -+ oo,
Loo n=O
1 ( 1 + (1- s)T) log(1- s) sn (mn(T) -log T) =--log -+ =-
T
1-s
316
1-s
Loo snan n=1
Solutions [6.15.34]-[6.15.35]
Problems
if lsi < 1, where an = ~?= 1 i- 1 , as required. (c) The mean velocity at time t is 1 aa 1 -+2 as s=l =t+-. 2 J
(0, 1)
(1, 0)
34. It is clear that Y is a Markov chain, and its possible transitions are illustrated in the above diagram. Let x andy be the probabilities of ever reaching (1, 1) from (1, 2) and (2, 1), respectively. By conditioning on the first step and using the translational symmetry, we see that x = ~ y + ~ x 2 andy= ~ + ~xy. Hence x 3 - 4x 2 + 4x- 1 = 0, an equation with roots x = 1, ~(3 ± 0). Since xis a probability, it must be that either x = 1 or x = ~(3- 0), with the corresponding values of y = 1 and y = ~ (0 - 1). Starting from any state to the right of (1 , 1) in the above diagram, we see by recursion that the chance of ever visiting (1, 1) is of the form xa yf3 for some non-negative integers a, {3. The minimal non-negative solution is therefore achieved when x = ~(3- 0) and y = ~(0- 1). Since x < 1, the chain is transient. 35. We write A, 1, 2, 3, 4, 5 for the vertices of the hexagon in clockwise order. Let Ti = min{n ::: 1: Xn = i} and ll"iO = ll"(· I Xo = i). (a) By symmetry, the probabilities Pi = ll"i (h < Tc) satisfy 2
PA = 3PI•
whence p A =
I PI= 3
I + 3P2·
I P2 = 3PI
1 + 3P3,
2
P3 = 3P2,
:J:r.
(b) By Exercise (6.4.6), the stationary distribution is nc = -1 8 JTA = . (c) By the argument leading to Lemma (6.4.5), this equals
:!, JTi
f-tAJTC
~ fori =/= C, whence MA =
= 2.
(d) We condition on the event E = {h < Tc} as in the solution to Exercise (6.2.7). The probabilities bi = lP'i (E) satisfy
317
[6.15.36]-[6.15.37]
Solutions
yielding bi = -ft, b2 = equations of the form
j;, b3
Markov chains
= ~- The transition probabilities conditional onE are now found by
ii2 =
li"2(E)p12 li"I (E)
. 'IarIy r2I = 97 , r23 = 92 , r32 = 2I , TIA = 76 . Hence, and sum IL2A = 1 +~ILIA+
giving ILIA=
~IL3A,
WI'th
IL3A = 1 + IL2A·
. . theobvwus notation, ILIA= 1 + +IL2A·
lj!, and the required answer is 1 +ILIA= 1 + .!j1 = J;f.
36. (a) We have that f3(m - i) 2
Pi,i+I =
m2
Look for a solution to the detailed balance equations
to find the stationary distribution
(b) In this case, f3(m - i)
a(i
Pi+I,i =
Pi,i+I =
m Look for a solution to the detailed balance equations f3(m - i)
lri
m
= lri+I
a(i
+ 1) m
+ 1) m
yielding the stationary distribution
37. We have that
by the Chapman-Kolmogorov equations
by the concavity of c
318
Solutions [6.15.38]-[6.15.41]
Problems
where we have used the fact that 'E,j njPjk(t) = nk. Now aj(s)---+ nj ass---+ oo, and therefore d(t) ---+ c(l).
38. By the Chapman-Kolmogorov equations and the reversibility, uo(2t) = LlP'(X(2t) = 0 I X(t) = j)lP'(X(t) = j
I X(O) = 0)
j
' (u·(t)) = "'no L..J ----:-1P'(X(2t) = j I X(t) = O)uj(t) =no " L..Jnj - 1 -. j
j
n]
2
n]
The function c(x) = - x 2 is concave, and the claim follows by the result of the previous problem.
39. This may be done in a variety of ways, by breaking up the distribution of a typical displacement and using the superposition theorem (6.13.5), by the colouring theorem (6.13.14), or by Renyi's theorem (6.13.17) as follows. Let B be a closed bounded region of Rd. We colour a point of IT at x E Rd black with probability lP'(x + X E B), where X is a typical displacement. By the colouring theorem, the number of black points has a Poisson distribution with parameter {
A.lP'(x + X E B) dx = A. {
}Jf!.d
=A.
lP'(X E dy - x)
dy {
}yEB
r
lxEJRd
dy
}yEB
r
lP'(X
E
dv) = A.IBI,
lvE!Rd
by the change of variables v = y - x. Therefore the probability that no displaced point lies in B is e-J..!BI, and the claim follows by Renyi's theorem.
40. Conditional on the number N (s) of points originally in the interval (0, s ), the positions of these points are jointly distributed as uniform random variables, so the mean number of these points which lie in ( -oo, a) after the perturbation satisfies A.s
looo Fx(a-u)du=IE(RL) lo0s1-lP'(X+u:sa)du---+A. s 0
as s ---+ oo,
where X is a typical displacement. Likewise, IE(LR) =A. J000 [1 - Fx(a + u)] du. Equality is valid if and only if
1
00
[1- Fx(v)]dv
=
a
ja-oo
Fx(v)dv,
which is equivalent to a =IE( X), by Exercise (4.3.5). The last part follows immediately on setting Xr = Vrt, where Vr is the velocity of the rth car.
41. Conditional on the number N(t) of arrivals by timet, the arrival times of these ants are distributed as independent random variables with the uniform distribution. Let U be a typical arrival time, so that U is uniformly distributed on (0, t). The arriving ant is in the pantry at time t with probability n = lP'(U +X > t), or in the sink with probability p = lP'(U +X < t < U +X+ Y), or departed with probability 1 - p - n. Thus, IE(xA(t)YB(t)) = IE{IE(xA(t)YB(t) I N(t))}
= IE{ (n x + py + 1 -
n - p )N(t)}
= eh(x-1) eJ..p(y-1).
Thus A(t) and B(t) are independent Poisson-distributed random variables. If the ants arrive in pairs and then separate,
319
[6.15.42]-[6.15.45]
Solutions
Markov chains
where y = 1- n- p. Hence,
whence A(t) and B(t) are not independent in this case.
42. The sequence {Xr} generates a Poisson process N(t) = max{n : Sn _:::: t}. The statement that Sn =tis equivalent to saying that there are n- 1 arrivals in (0, t), and in addition an arrival at t. By Theorem (6.12.7) or Theorem (6.13.11), the first n - 1 arrival times have the required distribution. Part (b) follows similarly, on noting that fu(u) depends on u = (UJ, u2, ... , un) only through the constraints on the Ur.
43. Let Y be a Markov chain independent of X, having the same transition matrix and such that Yo has the stationary distribution Jr. LetT= rnin{n ::: 1 : Xn = Yn} and suppose Xo = i. As in the proof of Theorem (6.4.17), IPij (n)- Tl:j I =
IL
nk(Pij (n)- Pkj (n))
k
I_: L
ll:klP'(T > n) = lP'(T > n).
k
Now,
lP'(T > r where
E
+1 IT
for r::: 0,
> r) _:::: 1- E2
= rnin;j {Pi}} > 0. The claim follows with "A = 1 -
E2 •
44. Let/k(n)betheindicatorfunctionofavisittokattimen,sothatE(h(n)) =lP'(Xn =k) =ak(n), say. By Problem (6.15.43), lak(n)- nkl _:::: ;..n. Now,
Lets = rnin{m, r} and t = lm- rl. The last summation equals
= n12
L L {(a;(s)- n;)(Pii (t)- n;) + n;(p;; (t)- n;) r
m
+ n; (a; (s) -
n;) - n; (a; (r) - n;) - n; (a; (m) - n;)}
1 _:: : 2 LL(;..s+t +"At +"As +"Ar +"Am) n r m as n--+ oo, where 0 < A < oo. For the last part, use the fact that I:~;:J f(Xr) = l:::iES f(i)V; (n). The result is obtained by Minkowski's inequality (Problem (4.14.27b)) and the first part.
45. We have by the Markov property that f(Xn+i I Xn, Xn-i, ... , Xo) = f(Xn+i I Xn), whence E(log f(Xn+i I Xn, Xn-J, ... , Xo) I Xn, ... , Xo) = E(log f(Xn+i I Xn) I Xn).
320
Problems
Solutions
[6.15.46]-[ 6.15.48]
Taking the expectation of each side gives the result. Furthermore, H(Xn+! I Xn) =- 2)Pij log Pij )lP'(Xn = i). i,j
Now X has a unique stationary distribution is finite, and the claim follows.
JC,
so that lP'(Xn = i) ---+ Jri as n ---+ oo. The state space
46. Let T = inf{t : X 1 = Y1 }. Since X and Y are persistent, and since each process moves by distance 1 at continuously distributed times, it is the case that lP'(T < oo) = 1. We define
Zt= {
Xt
if t < T,
Yt
ift 2: T,
noting that the processes X and Z have the same distributions. (a) By the above remarks, IIP'(Xt = k) -lP'(Yt = k)l = llP'(Zt = k) -lP'(Yt = k)l :::silP'CZt=k, T:::St)+lP'(Zt=k, T>t)-lP'(Yr=k, T:::St)-lP'(Yt=k, T>t)l :::S lP'(Xt = k, T > t)
+ lP'(Yt
= k, T > t).
We sum over k E A, and let t ---+ oo. (b) We have in this case that Z 1 :::S Y1 for all t. The claim follows from the fact that X and Z are processes with the same distributions. 47. We reformulate the problem in the following way. Suppose there are two containers, Wand N, containing n particles in all. During the time interval (t, t + dt), any particle in W moves toN with probability [tdt +o(dt), and any particle inN moves toW with probability A.dt +o(dt). The particles move independently of one another. The number Z(t) of particles in W has the same rules of evolution as the process X in the original problem. Now, Z (t) may be expressed as the sum of two independent random variables U and V, where U is bin(r, e1 ), Vis bin(n- r, 1/11), and e1 is the probability that a particle starting in W is in W at timet, 1/11 is the probability that a particle starting inN at 0 is in W at t. By considering the two-state Markov chain of Exercise (6.9.1),
and therefore E(sX(t)) = E(sU )E(s V) = (set
+1-
sr (s1flt
+1-
s)n-r.
Also, E(X(t)) = re 1 + (n- r)1/1 1 and var(X(t)) = re 1 (1- e1 ) + (n- r)1/f1 (1 -1/11 ). In the limit as n ---+ oo, the distribution of X (t) approaches the bin(n, A./(A. + ft)) distribution. 48. Solving the equations
gives the first claim. We have that y = l:i (Pi - qi )Jri, and the formula for y follows. Considering the three walks in order, we have that: A. Jri =
1for each i, and YA = -2a < 0.
B. Substitution in the formula for YB gives the numerator as 3{- ~a +o(a) }, which is negative for small a whereas the denominator is positive.
321
[6.15.49]-[6.15.51]
Solutions
Markov chains
i (Jb -
C. The transition probabilities are the averages of those for A and B, namely, Po = a)+ a) = a, and so on. The numerator in the formula for YC equals +o(l), which is positive for small a.
i
-fu-
-rfu
49. Call a car green if it satisfies the given condition. The chance that a green car arrives on the scene during the time interval (u, u +h) is A.hlP'(V < xf(t- u)) for u < t. Therefore, the arrival process of green cars is an inhomogeneous Poisson process with rate function A.(u) = {
A.lP'(V < xf(t- u)) 0
if u < t, if u :::: t.
Hence the required number has the Poisson distribution with mean A.
lot lP' ( V < t ~ J du lot lP' (v < ~) du lot E(/[Vu
= A.E(v- 1 min{x, Vt}).
50. The answer is the probability of exactly one arrival in the interval (s, t), which equals g(s) = A.(t- s)e-J..(t-s). By differentiation, g has its maximum atS = max{O, t - A.- 1}, and g(s) = e- 1 whent:::: A.- 1 . 51. We measure money in millions and time in hours. The number of available houses has the Poisson distribution with parameter 30A., whence the number A of affordable houses has the Poisson distribution with parameter~ ·30A. =SA. (cf. Exercise (3.5.2)). Since each viewing timeT has moment generating function E(e 8T) = (e 28 - e8 )je, the answer is
322
7 Convergence of random variables
7.1 Solutions. Introduction 1. (a) El(cXYI = lcr · {IIXIIr}'. (b) This is Minkowski's inequality. (c) Let E > 0. Certainly IX I 2: h where h is the indicator function of the event {I X I > E}. Hence EIX' I 2: Eli; I = lP'(IXI > E), implying that lP'(IXI > E) = 0 for all E > 0. The converse is trivial.
2.
(a) E({aX
+ bY}Z) = + E({X -
(b) E({X + Y} 2 ) (c) Clearly
aE(XZ)
+ bE(YZ). + 2E(Y 2 ).
Y} 2 ) = 2E(X 2 )
3. Let f(u) = ~E, g(u) = 0, h(u) = -~E, for all u. Then df(f, g)+ dE(g, h) = 0 whereas df(f, h)= 1. ~
Either argue directly, or as follows. With any distribution function F, we may associate a graph F obtained by adding to the graph of F vertical line segments connecting the two endpoints at each discontinuity ofF. By drawing a picture, you may see that .../2 d (F, G) equals the maximum distance between F and G measured along lines of slope -1. It is now clear that d(F, G) = 0 if and only if F = G, and that d(F, G) = d(G, F). Finally, by the triangle inequality for real numbers, we have that d(F, H) ~ d(F, G)+ d(G, H). 5.
Take X to be any random variable satisfying E(X 2 ) = oo, and define Xn = X for all n.
7.2 Solutions. Modes of convergence 1.
(a) By Minkowski's inequality,
let n --+ oo to obtain lim infn---+oo EIX~ I 2: EIX' 1. By another application ofMinkowski's inequality,
whencelimsupn 400 EIX~I ~ EIX'I.
323
[7.2.2]-[7.2.4]
Solutions
Convergence of random variables
(b) We have that IE(Xn)- E(X)I = IE(Xn- X) I ~ EIXn- XI--+ 0 as n --+ oo. The converse is clearly false. If each Xn takes the values ±1, each with probability then E(Xn) = 0, but EIXn - 01 = 1.
i,
(c) By part (a), E(X~) --+ E(X 2 ). Now Xn _..;. X by Theorem (7.2.3), and therefore E(Xn) --+ E(X) by part (b). Therefore var(Xn) = E(X~) - E(Xn) 2 --+ var(X).
2. Assume that Xn ~ X. Since IXnl ~ Z for all n, it is the case that lXI Zn = IXn -XI satisfies Zn ~ 2Z a.s. In addition, if E > 0,
~ Z a.s. Therefore
As n--+ oo, lP'(IZnl >E) --+ 0, and therefore the last term tends to 0; to see this, use the fact that E(Z) < oo, together with the result of Exercise (5.6.5). Now let E ,!.. 0 to obtain that EIZnl --+ 0 as n --+ oo. 3. We have that X- n- 1 ~ Xn ~ X, so that E(Xn) --+ E(X), and similarly E(Yn) --+ E(Y). By the independence of Xn and Yn,
E
{( 1) ( 1) } X - -;;_
Y - -;;_
= E(XY) -
E(X) + E(Y) 1 n + n 2 --+ E(XY)
as n --+ oo, so that E(Xn Yn) --+ E(XY).
4. Let F1, F2, ... be distribution functions. As in Section 5.9, we write Fn --+ F if Fn(x) --+ F(x) for all x at which F is continuous. We are required to prove that Fn --+ F if and only if d (Fn, F) --+ 0. Suppose that d(Fn, F) --+ 0. Then, forE > 0, there exists N such that F(x- E)- E ~ Fn(x)
~
F(x +E)+ E
for all x.
Take the limits as n --+ oo and E --+ 0 in that order, to find that Fn (x) --+ F (x) whenever F is continuous at x. Suppose that Fn --+ F. Let E > 0, and find real numbers a = x1 < x2 < · · · < Xn = b, each being points of continuity of F, such that (i) Fi(a) < E for all i, F(b) > 1- E, (ii) lxi+1 -Xi I < E for 1 ~ i < n. In order to pick a such that Fi (a) < E for all i, first choose a 1 such that F(a 1) < iE and F is continuous at a', then find M such that IFm (a')- F(a 1 ) I < iE form ::::: M, and lastly find a continuity point a ofF such that a~ a 1 and Fm(a) < E for 1 ~ m < M. There are finitely many points Xi, and therefore there exists N such that IFm(Xi)- F(xi)l < E for all i and m ::::: N. Now, if m ::::: N and Xi ~ x < Xi+ 1,
and similarly Fm(x)::::: Fm(Xi) > F(xi)- E :0:: F(x- E)- E.
Similar inequalities hold if x d(Fm, F) --+ 0 as m --+ oo.
~
a or x ::::: b, and it follows that d(Fm, F) < E if m ::::: N. Therefore
324
Modes of convergence 5.
n:::::
Solutions [7.2.5]-[7.2.7]
(a) Suppose c > 0 and pick 8 such that 0 < 8 < c. Find N such that lP'(/Yn - cJ > 8) < 8 for N. Now, for x::::: 0,
lP'(XnYn.::; x) .::; IP'(XnYn .::; X, /Yn- cJ .:'S
8) +
IP'(/Yn- c/ >
8)
.:'S lP' ( Xn .:'S c ~ 8 ) + 8,
and similarly
lP'(XnYn > x).::; IP'(XnYn >X, /Yn- cJ .:'S
8)
+ 8 .:'S lP' ( Xn > c: 8 ) + 8.
Taking the limits as n --+ oo and 8 .,),. 0, we find that lP'(XnYn .::; x) --+ lP'(X .::; xfc) if xfc is a point of continuity of the distribution function of X. A similar argument holds if x < 0, and we conclude that XnYn
~ eX if c > 0. No extra difficulty arises if c < 0, and the case c = 0 is similar.
For the second part, it suffices to prove that Y,;- 1 ~ c- 1 if Yn ~ c (# 0). This is immediate from the fact that IYn- 1 - c- 1 1 < E/{/c/(/c/- E)} if /Yn - c/ < E ( < Jcl). (b) Let E > 0. There exists N such that
lP'(/Xn/ >E)< E,
IP'(/Yn- Y/ >E)< E,
ifn
::0:: N,
and in addition lP'(/YI > N)
/g(x 1, y 1) - g(O, y)/ <
if Jx'J .::; 8, Jy'-
E
y/ .::; 8.
If /Xn/.::; 8, /Yn- Y/ .::; 8, and IYI .::; N, then Jg(Xn, Yn)- g(O, Y)J < E, so that
IP'(/g(Xn, Yn)- g(O, Y)J
::0:: E).::;
for n::::: N. Therefore g(Xn, Yn)
6.
lP'(/Xn/ > 8) + IP'(IYn- Y/ > 8) + lP'(/Y/ > N).::; 3E,
p ---+ g(O,
Y) as n--+ oo.
The subset A of the sample space Q may be expressed thus:
nU n 00
A=
00
00
{/Xn+m- Xn/ < k- 1 },
k=1 n=1m=1
a countable sequence of intersections and unions of events. For the last part, define
X(w) = {
limn---+oo Xn(w)
ifw E A
0
ifw
rf. A.
The function X is .1"-measurable since A E :F.
7. (a) If Xn(w)--+ X(w) then cnXn(w)--+ cX(w). (b) We have by Minkowski's inequality that, as n --+ oo,
E(/cnXn- cXJ').::; /cn/'E(/Xn- XI')+ /en- cJ'EJX'J--+ 0. (c) If c = 0, the claim is nearly obvious. Otherwise c # 0, and we may assume that c > 0. For 0 < E < c, there exists N such that /en - cJ < E whenever n ::::: N. By the triangle inequality, /cnXn- eX/ .::; /cn(Xn- X)/+ /(en- c)X/, so that, for n ::0:: N,
IP'(/cnXn- eX/> E).::; IP'(cn/Xn- X/> iE) + IP'(/cn- cJ·JXJ > iE) .:'S lP' (/Xn -X/ > --+ 0
E ) + lP' (/X/ > E ) 2(c +E) 2/cn - cJ as n --+ oo.
325
[7.2.8]-[7.3.1]
Solutions
Convergence of random variables
_!;. X, find random variables ~ cY, so that cnYn _!;. cY,
(d) A neat way is to use the Skorokhod representation (7.2.14). If Xn Yn, Y with the same distributions such that Yn implying the same conclusion for the X's.
8.
~
Y. Then cnYn
If X is not a.s. constant, there exist real numbers c and E such that 0 < E <
lP'(X > c +E) > 2E. Since Xn
i and lP'( X < c) > 2E,
!.. X, there exists N such that
lP'(Xn
lP'(Xn>c+E)>E,
ifn;:::N.
Also, by the triangle inequality, IXr- Xsl _:::: IXr- XI+ IXs- XI; therefore there exists M such that IP'(IX, - Xs I > E) < E3 for r, s ::::: M. Assume now that the Xn are independent. Then, for r, s ::::: max{M, N}, r =f. s, E3 > IP'(IXr - Xs I > E)
IP'(Xr < c, Xs > c +E) = lP'(Xr < c)lP'(Xs > c +E) > E2 ,
::0::
a contradiction. 9. Either use the fact (Exercise (4.12.3)) that convergence in total variation implies convergence in distribution, together with Theorem (7.2.19), or argue directly thus. Since luOI :::: K < oo, IE(u(Xn))- E(u(X))I = jL:u(k){fn(k)- /(k)}l _:::: KL lfn(k)- /(k)l--+ 0. k
10. The partial sum Sn
= 2::~= 1
k
X, is Poisson-distributed with parameter an
= 2::~= 1 A.,.
For fixed
x, the event {Sn _:::: x} is decreasing inn, whence by Lemma (1.3.5), if an --+ a < oo and xis a non-negative integer, oo
lP' ( LXr r=1
)
:::=x =n~~lP'(Sn ::;x)
=
x
-crj
)=0
} ·
L~·
Hence if a < oo, 2::~ 1 X, converges to a Poisson random variable. On the other hand, if an --+ oo,
L:J=
then e-crn 0 aj I j! --+ 0, giving that IP'(2::~ 1 X, > x) diverges with probability 1, as required.
=
I for all
x,
and therefore the sum
7.3 Solutions. Some ancillary results 1.
(a) If IXn - Xm I > E then either IXn- XI > iE or IXm -XI > iE, so that
as n, m --+ oo, forE > 0. Conversely, suppose that {Xn} is Cauchy convergent in probability. For each positive integer k, there exists nk such that forn,
m::::: nk.
The sequence (nk) may not be increasing, and we work instead with the sequence defined by N1 = n1, Nk+1 = max{Nk + 1, nk+d· We have that LlP'(IXNk+J- XNkl::::: Tk) < k
326
00,
Solutions
Some ancillary results
[7.3.2]-[7.3.3]
whence, by the first Borel-Cantelli lemma, a.s. only finitely many of the events {IXNk+l - XNk I 2:: 2-k} occur. Therefore, the expression 00
X= XN 1 + L(XNk+i - XNk) k=1 converges absolutely on an event C having probability one. Define X(w) accordingly for wE C, and X(w) = 0 for w rf. C. We have, by the definition of X, that XNk ~ X ask-+ oo. Finally, we 'fill in the gaps'. As before, forE > 0,
as n, k-+ oo, where we are using the assumption that {Xn} is Cauchy convergent in probability. (b) Since Xn
!.. X, the sequence {Xn} is Cauchy convergent in probability. Hence IP'(IYn- Yml >E)= IP'(IXn- Xml >E)-+ 0
as n, m-+ oo,
forE > 0. Therefore {Yn} is Cauchy convergent also, and the sequence converges in probability to some limit Y. Finally, Xn 2.
.£.. X and Yn .£.. X, so that X andY have the same distribution.
Since An ~ U;;'=n Am, we have that
lJ
lJ
limsuplP'(An) _:::: lim lP'( Am) = lP'( lim Am) = lP'(An i.o.), n--+oo n--+oo m=n n--+oo m=n where we have used the continuity of lP'. Alternatively, apply Fatou's lemma to the sequence (4 ~ of indicator functions. 3. (a) Suppose Xzn = 1, Xzn+1 = -1, for n 2:: 1. Then {Sn = 0 i.o.} occurs if X 1 = -1, and not if X 1 = 1. The event is therefore not in the tail a-field of the X's. (b) Here is a way. As usual, lP'(Szn = 0) =
e:)
{p(l - p W, so that
LlP'(Szn = 0) < oo
if p
=f.
i.
n
implying by the first Borel-Cantelli lemma that lP'(Sn = 0 i.o.) = 0. (c) Changing the values of any finite collection of the steps has no effect on I = lim inf Tn and J = lim sup Tn, since such changes are extinguished in the limit by the denominator '.[ii'. Hence I and J are tail functions, and are measurable with respectto the tail a -field. In particular, {I _:::: - x} n {J ::::: x} lies in the a -field. Take x = 1, say. Then, lP'(/ _:::: -1) = lP'(J 2:: 1) by symmetry; using Exercise (7.3.2) and the central limit theorem, lP'(J 2:: 1) 2:: lP'(Sn 2:: .[ii i.o.) 2:: lim sup lP'(Sn 2:: n--+oo
,Jfi,) =
1 - (l) > 0,
where is theN (0, 1) distribution function. Since {J 2:: 1} is a tail event of an independent sequence, it has probability either 0 or 1, and therefore lP'(/ _:::: -1) = lP'(J 2:: 1) = 1, and also lP'(/ _:::: -1, J 2:: 1) = 1. That is, on an event having probability one, each visit of the walk to the left of -.[ii is followed by a visit of the walk to the right of .[ii, and vice versa. It follows that the walk visits 0 infinitely often, with probability one.
327
[7.3.4]-[7.3.8]
Solutions
Convergence of random variables
4. Let A be exchangeable. Since A is defined in terms of the Xi, it follows by a standard result of measure theory that, foreachn, thereexistsaneventAn E a(X1, X2, ... , Xn), suchthatlP'(Ab.An)--+ 0 as n --+ oo. We may express An and A in the form An = {Xn E Bn},
where Xn
A = {X E B},
= (X 1, X2, ... , Xn), and Bn and Bare appropriate subsets ofll~n and JR 00 • Let A~ ={X~
E
Bn},
A'= {X 1 E B},
where X~ = (Xn+1, Xn+2, ... , X2n) and X'= (Xn+1, Xn+2, ... , X2n, X1, X2, ... , Xn, X2n+1, X2n+2, · · · ). Now !P'(An n A~) = lP'(An)lP'(A~), by independence. Also, !P'(An) = lP'(A~), and therefore
!P'(A
By the exchangeability of A, we have that lP'(A t. A~) = !P'(A' t. A~), which in turn equals t. An), using the fact that the Xi are independent and identically distributed. Therefore, liP'( An
n A~)
- lP'(A) I ::: lP'(A
t.
An)
+ !P'(A t. A~) --+ 0
as n --+ oo.
Combining this with(*), we obtain that lP'(A) = lP'(A) 2, and hence lP'(A) equals 0 or 1.
5. The value of Sn does not depend on the order of the first n steps, but only on their sum. If Sn = 0 i.o., then S~ = 0 i.o. for all walks {S~} obtained from {Sn} by permutations of finitely many steps. 6. Since f is continuous on a closed interval, it is bounded: If (y) I ::: c for all y E [0, 1]for some c. Furthermore f is uniformly continuous on [0, 1], which is to say that, if E > 0, there exists 8 (> 0), such that lf(y)- f(z)l < E if IY- zl :S 8. With this choice of E, 8, we have that IE(Z/Ac)l < E, and IE(ZIA)I ::: 2clP'(A) ::: 2c ·
x(1 - x)
n 82
by Chebyshov's inequality. Therefore 2c IE(Z)I < E + n 82 , which is less than 2E for values of n exceeding 2c/(E8 2).
7. If {Xn} converges completely to X then, by the first Borel-Cantelli lemma, IXn -XI > E only finitely often with probability one, for all E > 0. This implies that Xn ~ X; see Theorem (7.2.4c). Suppose conversely that {Xn} is a sequence of independent variables which converges almost surely to X. By Exercise (7.2.8), X is almost surely constant, and we may therefore suppose that Xn ~ c where c E JR. It follows that, for E > 0, only finitely many of the (independent) events {IXn - cl > E} occur, with probability one. Using the second Borel-Cantelli lemma, LlP'(IXn- cl >E)<
00.
n
8.
Of the various ways of doing this, here is one. We have that
(n) 2
-1
" ' XiXj L · · 1~l
=
_n (_!_ ~
n
-
1 n L z"--1
328
xi)2-
~
1 x2 n(n - 1) L. 1 • z--1
Solutions [7.3.9]-[7.3.12]
Some ancillary results
Now n
-1
L;'i Xi
D
-+ f.l, by the law oflarge numbers (5.10.2); hence
P n -1 L;'i Xi -+ ~t (use Theorem
(7.2.4a)). It follows that (n- 1 L;'i Xi) 2 ~ ~t 2 ; to see this, either argue directly or use Problem (7.11.3). Now use Exercise (7.2.7) to find that n ( 1 n --I: xi )2 -+p ~t 2 . n- 1 n i= 1
Arguing similarly,
1
n
- - - """'x? ~ o n(n- 1) ~ !=1
'
'
and the result follows by the fact (Theorem (7.3.9)) that the sum of these two expressions converges in probability to the sum of their limits. 9.
Evidently,
Xn ) = -1lP' ( -->1+E logn nl+E'
for lEI < 1.
By the Borel-Cantelli lemmas, the events An = {Xn/logn -1 < E ::=: 0, and a.s. only finitely often for E > 0.
~
1 + E} occur a.s. infinitely often for
10. (a) Mills's ratio (Exercise (4.4.8) or Problem (4.14.1c)) informs us that 1-
1
.J2n logn(l + E)nO+E)
2.
The sum over n of these terms converges if and only if E > 0, and the Borel--Cantelli lemmas imply the claim. (b) This is an easy implication of the Borel--Cantelli lemmas. 11. Let X be uniformly distributed on the interval [-1, 1], and define Xn = /{X:S(-l)n /n)· The distribution of Xn approaches the Bernoulli distribution which takes the values ± 1 with equal probability The median of X n is 1 if n is even and - 1 if n is odd.
i.
12. (i) We have that
for x > 0. The result follows by the second Borel--Cantelli lemma. (ii) (a) The stationary distribution 1l is found in the usual way to satisfy
Hence nk = {k(k + 1)}- 1 fork~ 1, a distribution with mean L;~ 1 (k + 1)- 1 (b) By construction, lP'(Xn ::=: Xo + n) = 1 for all n, whence
. sup -Xn ::=: 1) lP' ( hm n-+oo n It may in fact be shown that lP'(lim supn-+oo Xn/n
=
= 0) = 1.
329
1.
= oo.
[7 .3.13]-[7.4.3]
Solutions
Convergence of random variables
13. We divide the numerator and denominator by Jna. By the central limit theorem, the former converges in distribution to the N(O, 1) distribution. We expand the new denominator, squared, as
-
12:n
na 2
(Xr -
~-t)
2
2-
- -(Xna 2
r=1
~-t)
I:n (Xr
-
~-t)
1-
+-(Xa2
r=1
~-t)
2
.
By the weak law of large numbers (Theorem (5.10.2), combined with Theorem (7.2.3)), the first term converges in probability to 1, and the other terms to 0. Their sum converges to 1, by Theorem (7.3.9), and the result follows by Slutsky's theorem, Exercise (7.2.5).
7.4 Solutions. Laws of large numbers 1.
LetSn=X1+X2+···+Xn.Then n
.
2
'"'L 1 -< nE(S 2) = n i= 2 log i - log n
and therefore Snfn ~ 0. On the other hand, I::i lP'(IXil 2: i) = 1, so that IXil 2: i i.o., with probability one, by the second Borel-Cantelli lemma. For such a value of i, we have that ISi - Si -11 2: i, implying that Sn/ n does not converge, with probability one.
2.
Let the Xn satisfy
1 n
lP'(Xn = -n) = 1 - 2•
lP'(Xn
= n3 -
n)
= 21 , n
whence they have zero mean. However,
implying by the first Borel-Cantelli lemma that lP'(Xnfn-+ -1) = 1. It is an elementary result of real analysis that n - 1 I::~= 1 Xn -+ -1 if Xn -+ -1, and the claim follows. 3. The random variable N(S) has mean and variance AlSI = crd, where cis a constant depending only on d. By Chebyshov's inequality,
lP'
(I
I )
N(S) ISl-A :::
E
A (A) ~
.:5 E21SI =
By the first Borel-Cantelli lemma, I1Ski- 1N(Sk)- AI 2:
E
2
era·
for only finitely many integers k, a.s.,
where Skis the sphere of radius k. It follows that N(Sk)/ISkl ~ A ask-+ oo. The same conclusion holds ask -+ oo through the reals, since N(S) is non-decreasing in the radius of S.
330
Solutions
Martingales
[7.5.1]-[7.7.1]
7.5 Solutions. The strong law 1.
Let Iij be the indicator function of the event that X J lies in the i th interval. Then n
log Rm
=L
n
Zm (i) log Pi
=L
i=l
where, for 1 .::0 j .::0 m, Yj variables with mean
m
L
m
lij log Pi
=L
i=l }=I
l:t=I Iij
Yj
}=I
log Pi is the sum of independent identically distributed n
lE(Yj) =
L Pi log Pi = -h. i=l
By the strong law, m- 1 log Rm ~ -h. 2. The following two observations are clear: (a) N(t) < n if and only if Tn > t, (b) TN(t) .::0 t < TN(t)+I· If lE(X I) < oo, then lE(Tn) < oo, so that lP'(Tn > t) -+ 0 as t -+ oo. Therefore, by (a), lP'(N(t) < n) = lP'(Tn > t) -+ 0
as t-+ oo,
implying that N(t) ~ oo as t-+ oo. Secondly, by (b), TN(t) < _t_ < TN(t)+I . ( 1 + N(t)-I). N(t) - N(t) N(t) + 1 Take the limit as t -+ oo, using the fact that Tn/n ~ lE(XJ) by the strong law, to deduce that t/N(t) ~ lE(XJ).
By the strong law, Sn/n ~ lE(XJ) =I= 0. In particular, with probability 1, Sn = 0 only finitely often.
3.
7.6 Solution. The law of the iterated logarithm 1.
The sum Sn is approximately N(O, n), so that -~a
lP'(Sn > ylanlogn) = 1-
~ alogn
for all large n, by the tail estimate of Exercise (4.4.8) or Problem (4.14.1c) for the normal distribution. This is summable if a > 2, and the claim follows by an application of the first Borel-Cantelli lemma.
7.7 Solutions. Martingales 1.
Suppose i < j. Then
_I]} = lE{ Xi [lE(SJ I So, S1, ... , SJ-I)- s1_I]} = 0
lE(XjXi) = lE{ lE[(SJ- Sj-I)Xi
331
I So, S1, ... , s1
[7. 7.2]-[7.8.2]
Solutions
Convergence of random variables
by the martingale property. 2.
Clearly EISnl < oo for all n. Also, for n 2: 0,
1 { ( 1 _ p,n+1)} E(Sn+1 I Zo, z1' ... ' Zn) = p,n+1 E(Zn+1 I Zo, ... ' Zn)- m 1 - JL 1 { ( 1 _ p,n+1)} = p,n+ 1 m+p,Zn-m =Sn. 1 _p, 3.
CertainlyEISnl < ooforalln. Secondly,forn 2:1, E(Sn+1 I Xo, X1, ... , Xn) = aE(Xn+1 I Xo, ... , Xn) = (aa + 1)Xn +abXn-1,
+ Xn
which equals Sn if a= (1 - a)- 1. 4. The gambler stakes Zi = /i-1 (X 1, ... , Xi-1) on the ith play, at a return of Xi per unit. Therefore Si = Si-1 + XiZi fori 2: 2, with S1 = X1Y. Secondly,
where we have used the fact that Zn+1 depends only on X1, X2, ... , Xn.
7.8 Solutions. Martingale convergence theorem 1. It is easily checked that Sn defines a martingale with respect to itself, and the claim follows from the Doob-Kolmogorov inequality, using the fact that n
E(S~) = I:var(Xj). }=1
2. It would be easy but somewhat perverse to use the martingale convergence theorem, and so we give a direct proof based on Kolmogorov 's inequality of Exercise (7 .8 .1 ). Applying this inequality to the sequence Zm, Zm+1' ... where Zi =(Xi- EXi)/i, we obtain that Sn = Z1 + Z2 + · · · + Zn satisfies, for E > 0, lP' ( max ISn- Sml m
~ ~
>E) ::0 _!_2 E
var(Zn).
n=m+1
We take the limit as r -+ oo, using the continuity oflP', to obtain
1 00 1 lP' ( sup ISn- Sml > E ) ::0 2 2 var(Xn). n:=::m E n=m+1 n
.2:::
Now let m -+ oo to obtain (after a small step) lP' ( lim sup ISn - Sm I ::0 m-;.oo n:=::m
E) = 1
for all
E
Any real sequence (xn) satisfying lim sup lxn -
m-;.oo n:=::m
Xm
I ::0
332
E
for all
E
> 0,
> 0.
Solutions
Prediction and conditional expectation
[7.8.3]-[7.9.4]
is Cauchy convergent, and hence convergent. It follows that Sn converges a.s. to some limit Y. The last part is an immediate consequence, using Kronecker's lemma. 3. By the martingale convergence theorem, S = limn-+oo Sn exists a.s., and Sn Exercise (7.2.1c), var(Sn) --+ var(S), and therefore var(S) = 0.
m.s.
~
S. Using
7.9 Solutions. Prediction and conditional expectation 1. (a) Clearly the best predictors are lE(X (b) We have, after expansion, that
I Y) = Y 2 , lE(Y I X) = 0.
since lE(Y) = lE(Y 3 ) = 0. This is a minimum when b = JE(Y 2 ) = ~.and a = 0. The best linear predictor of X given Y is therefore ~ . Note that lE(Y I X) = 0 is a linear function of X; it is therefore the best linear predictor of Y given X. 2.
By the result of Problem (4.14.13), lE(Y I X) =
3.
Write
~-tz
+ paz(X -
~-tdfal,
in the natural notation.
n
g(a) = LaiXi =aX', i=l
and v(a) = lE{ (Y- g(a)) 2 } = lE(Y 2 ) - 2alE(YX1) + aVa'. Let a be a vector satisfying Vi' = lE(YX'). Then v(a) - v(a) = aVa' - 2alE(YX1) + 2ilE(YX1) -iVa' = aVa'- 2aVa' +iVa'= (a- a)V(a- a)' :::: 0, since Vis non-negative definite. Hence v(a) is a minimum when a =a, and the answer is g(i). If V is non-singular, a= JE(YX)V- 1 . 4. Recall that Z = JE(Y I fj,) is the ('almost') unique fJ,-measurable random variable with finite mean and satisfying JE{(Y- Z)/G} = 0 for all G E fJ,.
(i) Q E fJ,, and hence lE{lE(Y I fj,)/n} = lE(Z/n) = lE(Y /g). (ii) U = alE(Y I fj,) + ,BlE(Z I fj,) satisfies JE(U /G) = alE{lE(Y I fj,)/G} + ,BlE{lE(Z I fj,)/G} = alE(Y I G)+ ,BlE(Z/G) = lE{ (aY + ,BZ)IG }, Also, U is fJ,-measurable. (iii) Suppose there exists m (> 0) such that G = {lE(Y I fj,) < -m} has strictly positive probability. Then G E fJ,, and so JE(Y /G) = lE{lE(Y I fj,)/G}. However Y IG :::: 0, whereas lE(Y I fj,)/G < -m. We obtain a contradiction on taking expectations. (iv) Just check the definition of conditional expectation. (v) If Y is independent of fJ,, then lE(Y /G)= lE(Y)lP'(G) for G E fJ,. Hence JE{(Y -JE(Y))/G} = 0 for G E fJ,, as required.
333
[7.9.5]-[7.10.1]
Solutions
Convergence of random variables
(vi) If g is convex then, for all a E JR, there exists A.(a) such that g(y) 2: g(a)
+ (y- a)A.(a);
furthermore A. may be chosen to be a measurable function of a. Set a = E(Y I fJ,) andy = Y, to obtain g(Y) ::: g{E(Y I fJ,)} + {Y - E(Y I fJ,) }A.{E(Y I fJ,) }. Take expectations conditional on g,, and use the fact that E(Y I fJ,) is [J,-measurable. (vii) We have that IECYn I fJ.) -E(Y I fJ.)I :':: E{IYn- Yll fJ.} :':: Vn where Vn = E{supm;c:n IYm- Yll fJ.}. Now {Vn : n::: 1} is non-increasing and bounded below. Hence V = limn--+oo Vn exists and satisfies V ::: 0. Also E(V) :<:: E(Vn) = E {sup IYm- Yl}, m?:_n
which tends to 0 as m --+ oo, by the dominated convergence theorem. Therefore E(V) = 0, and hence lP'(V = 0) = 1. The claim follows.
5.
E(Y
I X)=
X.
6. (a) Let {Xn : n ::: 1} be a sequence of members of H which is Cauchy convergent in mean square, that is, E{IXn - Xm 12 } --+ 0 as m, n --+ oo. By Chebyshov's inequality, {Xn : n ::: 1} is Cauchy convergent in probability, and therefore converges in probability to some limit X (see Exercise (7.3.1)). It follows that there exists a subsequence {Xnk : k ::: 1} which converges to X almost surely. Since each Xnk is fJ,-measurable, we may assume that X is fJ,-measurable. Now, as n --+ oo,
where we have used Fatou's lemma and Cauchy convergence in mean square. Therefore Xn ~ X. That E(X 2 ) < oo is a consequence of Exercise (7.2.la). (b) That (i) implies (ii) is obvious, since I G E H. Suppose that (ii) holds. Any Z ( E H) may be written as the limit, as n --+ oo, of random variables of the form m(n) Zn =
L
a;(n)Ic;(n)
i=l
for reals a;(n) and events G;(n) in fJ,; furthermore we may assume that IZnl :<:: IZI. It is easy to see that E{ (Y- M)Zn} = 0 for all n. By dominated convergence, E{(Y- M)Zn} --+ E{(Y- M)Z}, and the claim follows.
7.10 Solutions. Uniform integrability 1.
It is easily checked by considering whether lxl :<::a or IYI :<::a that, for a > 0,
Now substitute x = Xn andy= Yn, and take expectations.
334
Solutions
Uniform integrability
[7.10.2]-[7.10.6]
xn
2. (a) Let E > 0. There exists N such that lE(IXn < E if n > N. Now lEIXrl < oo, by Exercise (7.2.1a), and therefore there exists 8 (> 0) such that lE(IXnlr lA) < E for 1 S n S N,
lE(IXIr lA) < E,
for all events A such that lP'(A) < 8. By Minkowski's inequality, {lE(IXn(/A)} 1/r S {lE(IXn- X(/A)} 1/r
+ {lE(IX(/A)} 1/r
S 2El/r
ifn > N
iflP'(A) < 8. Therefore {IXnr: n 2: 1} is uniformly integrable. If r is an integer then {X~ : n 2: 1} is uniformly integrable also. Also X~ ~ xr since Xn ~ X (use the result of Problem (7.11.3)). Therefore lE(X~) --+ E(Xr) as required. (b) Suppose now that the collection {IXnlr : n 2: 1} is uniformly integrable and Xn ~X. We show first that lEIXr I < oo, as follows. There exists a subsequence {Xnk : k 2: 1} which converges to X almost surely. By Fatou's lemma, lEIXrl = lE (liminf IXnklr) S k-H>O
liminflEIX~kl S suplEIX~I k-+oo n
< oo.
If E > 0, there exists 8 (> 0) such that lE(IXriiA) < E,
lE(IX~IIA) < E
for all n,
whenever A is such that lP'(A) < 8. There exists N such that Bn(E) = {IXn- XI > E} satisfies lP'(Bn(E)) < 8 for n > N. Consequently lE(IXn-
xn S
Er
+ E (IXn- Xlr IBn(E)),
n > N,
of which the final term satisfies {lE (IXn - Xlr IBn(E))} lfr S {lE (IX~IIBn(E))} lfr
+ {lE (IXr IIBn(E))} l/r
S 2El/r.
Therefore, Xn ~X. 3.
FixE > 0, and find a real number a such that g(x) > xjE if x >a. If b 2: a, lE (IXnii(IXnl>bJ) < ElE{g(IXnl)} S EsupE{g(IXnl)}, n
whence the left side approaches 0, uniformly inn, as b --+ oo.
4.
Here is a quick way. Extinction is (almost) certain for such a branching process, so that Zn ~ 0,
and hence Zn ~ 0. If {Zn : n 2: 0} were uniformly integrable, it would follow that JE(Zn) --+ 0 as n --+ oo; however lE(Zn) = 1 for all n. 5.
We may suppose that Xn, Yn, and Zn have finite means, for all n. We have that 0 S Yn - Xn S p
Zn- Xn where, by Theorem (7.3.9c), Zn- Xn-+ Z- X. Also
lEIZn- Xn I = lE(Zn- Xn)--+ lE(Z- X)
= lEIZ- XI,
so that {Zn - Xn : n 2: 1} is uniformly integrable, by Theorem (7.10.3). It follows that {Yn - Xn : n 2: 1} is uniformly integrable. Also Yn - Xn ~ Y- X, and therefore by Theorem (7.10.3), lEIYn-Xnl-+ lEIY-XI,whichistosaythatlE(Yn)-lE(Xn)--+ lE(Y)-lE(X);hencelE(Yn)--+ E(Y). It is not necessary to use uniform integrability; try doing it using the 'more primitive' Fatou's lemma.
6. For any event A, lE(IXniiA) S JE(ZIA) where Z = supn IXnl· The uniform integrability follows by the assumption that lE(Z) < oo.
335
[7.11.1]-[7.11.2]
Solutions
Convergence of random variables
7.11 Solutions to problems 1.
lEI X~ I = oo for r 2: 1, so there is no convergence in any mean. However, if E > 0, 2
lP'(IXnl >E)= 1 - - tan- 1 (nE)-+ 0
as n-+ oo,
T(
p
so that Xn -+ 0. You have insufficient information to decide whether or not Xn converges almost surely: (a) Let X be Cauchy, and let Xn = Xjn. Then Xn has the required density function, and Xn ~ 0. (b) Let the Xn be independent with the specified density functions. ForE > 0,
2 sin- 1 ( lP'(IXnl >E)=n
1
,h + n2E2
) "'2, nnE
so that Ln lP'(IXnl >E)= oo. By the second Borel-Cantelli lemma, IXnl > E i.o. with probability one, implying that Xn does not converge a.s. to 0. 2. (i) Assume all the random variables are defined on the same probability space; otherwise it is meaningless to add them together. (a) Clearly Xn(w) + Yn(w) -+ X(w) + Y(w) whenever Xn(w) -+ X(w) and Yn(w) -+ Y(w). Therefore { Xn + Yn...,. X+ Y} ~ {Xn...,. X} U {Yn...,. Y}, a union of events having zero probability. (b) Use Minkowski's inequality to obtain that {lE (IXn + Yn -X- Yn} 1/r .::0 {lE(IXn -
xn} 1/r +
{lE(IYn - Yn} 1/r.
(c) If E > 0, we have that
and the probability of the right side tends to 0 as n -+ oo. (d) If Xn ~ X and the Xn are symmetric, then -Xn ~ X. However Xn + (-Xn) ~ 0, which generally differs from 2X in distribution. (ii) (e) Almost-sure convergence follows as in (a) above. (f) The corresponding statement for convergence in rth mean is false in general. Find a random variable Z such that lEIZrl < oo but lEIZ 2rl = oo, and define Xn = Yn = Z for all n. p
p
(g) Suppose Xn -+ X and Yn -+ Y. Let E > 0. Then
lP'(IXnYn- XYI >E)= lP'(I(Xn- X)(Yn- Y) + (Xn- X)Y + X(Yn-
nl >E)
.::0 lP'(IXn- XI·IYn- Yl >~E) +lP'(IXn- XI·IYI >~E)
+ lP'(IXI ·IYn- Yl > ~E). Now, for 8 > 0, lP'(IXn- XI·IYI >~E) .::0 lP'(IXn- XI> E/(38)) +lP'(IYI >
336
8),
Solutions
Problems
[7 .11.3]-[7.11.5]
which tends to 0 in the limit as n --+ oo and 8 --+ oo in that order. Together with two similar facts, we obtain that XnYn ~ XY. (h) The example of (d) above indicates that the corresponding statement is false for convergence in distribution. 3. Let E > 0. We may pick M such that lP'(IXI 2: M) .::::E. The continuous function g is uniformly continuous on the bounded interval [- M, M]. There exists 8 > 0 such that lg(x)- g(y)l If lg(Xn)- g(X)I >
E,
.:::= E
.:::=
8 and lxl
.:::=
M.
then either IXn- XI > 8 or lXI 2: M. Therefore
lP'(Ig(Xn)- g(X)I
>E) _:: : lP'(IXn- XI > 8)
in the limit as n --+ oo. It follows that g(Xn)
4.
if lx- Yl
~
+ JI»(IXI 2:
M)
--+ lP'(IXI 2:
M) _:: : E,
g(X).
Clearly lE(eitXn)
n . n { 1 1 =II lE(eitYj/101) =II _. - e . . . . 10 1 _ eztf10J
it/10j-l }
]=1
-
]=1
1- eit wn(l-eitjlOn)
1 - eit --+----
it
as n --+ oo. The limit is the characteristic function of the uniform distribution on [0, 1]. Now Xn .::; Xn+1 _:::: 1 for all n, so that Y(w) = limn---+oo Xn(w) exists for all w. Therefore Xn
~
5.
(a) Supposes < t. Then
Y; hence Xn
Er Y, whence Y has the uniform distribution.
lE(N(s)N(t)) = JE(N(s) 2) +lE{N(s)(N(t)- N(s))} = JE(N(s) 2) +lE(N(s))lE(N(t)- N(s)), since N has independent increments. Therefore cov(N(s), N(t)) = lE(N(s)N(t)) -JE(N(s))JE(N(t)) = (A.s) 2 + A.s + A.s{A.(t- s)}- (A.s)(A.t) = A.s. In general, cov(N(s), N(t)) =A. min{s, t}. (b) N(t +h)- N(t) has the same distribution as N(h), if h > 0. Hence
which tends to 0 as h --+ 0. (c) By Markov's inequality,
lP'(IN(t+h)-N(t)l
>E).:::: E12lE({N(t+h)-N(t)} 2),
which tends to 0 as h --+ 0, if E > 0. (d) Let E > 0. For 0 < h < E- 1, lP'
(I
N(t +
I
h~- N(t) >E)
= lP'(N(t +h)- N(t):::
337
1)
= A.h + o(h),
[7.11.6]-[7.11.10]
Solutions
Convergence of random variables
which tends to 0 as h --+ 0. On the other hand,
which tends to oo as h -J, 0.
6.
By Markov's inequality, Sn
= 2:7= 1 X;
satisfies
forE > 0. Using the properties of the X's,
since E(X;) = 0 for all i. Therefore there exists a constant C such that
implying (via the first Borel--Cantelli lemma) that n- 1 Sn ~ 0.
7.
We have by Markov's inequality that <
00
forE > 0, so that Xn ~ X (via the first Borel-Cantelli lemma).
8. Either use the Skorokhod representation or characteristic functions. Following the latter route, the characteristic function of aXn + b is
where ¢>n is the characteristic function of Xn. The result follows by the continuity theorem.
9.
For any positive reals c, t, li"(X > t) -
= li"(X + c >- t + c)
< -
E{(X+c) 2 } . (t +c) 2
Set c = a- 2 jt to obtain the required inequality.
10. Note that g(u) = uj(l + u) is an increasing function on [0, oo). Therefore, forE > 0, lP' X > (I nl
E
)
=11" (
IXnl l+E > -E- ) < - .JE: ( IXnl ) l+IXnl l+E E l+IXnl
338
Solutions [7.11.11]-[7.11.14]
Problems
by Markov's inequality. If this expectation tends to 0 then Xn
~ 0.
p
Suppose conversely that Xn -+ 0. Then E(
IXnl ) E E :':: - - ·li"(IXnl :'::E)+ 1 ·lP'(IXnl >E)--+-1+1Xnl 1+E 1+E
as n --+ oo, for E > 0. However E is arbitrary, and hence the expectation has limit 0.
11. (i) The argument of the solution to Exercise (7.9.6a) shows that {Xn} converges in mean square if it is mean-square Cauchy convergent. Conversely, suppose that Xn ~ X. By Minkowski's inequality,
as m, n--+ oo, so that {Xn} is mean-square Cauchy convergent. (ii) The corresponding result is valid for convergence almost surely, in rth mean, and in probability. For a.s. convergence, it is self-evident by the properties of Cauchy-convergent sequences of real numbers. For convergence in probability, see Exercise (7.3.1). For convergence in rth mean (r ~ 1), just adapt the argument of (i) above.
12. Ifvar(Xi)::: M for all i, the variance of n- 1 Lt=1
xi is
M
1 n
2 L:var(Xi) :=:---+ 0 n i=1 n
as n--+ oo.
13. (a) We have that JP>(Mn ::0 GnX) = F(anx)n --+ H(x)
If x :=: 0 then F (an X)n --+ 0, so that H (x)
= 0.
as n--+
00.
Suppose that x > 0. Then
-logH(x) =- lim {nlog[1- (1- F(anx))]} = n----+-oo
lim {n(1- F(anx))}
n---+oo
since -y- 1 log(1 - y) --+ 1 as y t 0. Setting x = 1, we obtain n(l - F(an)) --+ -log H(l), and the second limit follows. (b) This is immediate from the fact that it is valid for all sequences {an}. (c) We have that
1- F(tex+y) 1 - F(t)
1- F(tex+y) 1- F(tex)
1-F(tex)
1 - F(t)
logH(eY)
--+ --=--='---:::c=:'-:-:-c'log H(1)
logH(ex)
log H(l)
as t --+ oo. Therefore g(x + y) = g(x)g(y). Now g is non-increasing with g(O) = 1. Therefore g(x) = e-f3x for some ,B, and hence H(u) = exp( -au-f3) for u > 0, where a= -log H(l).
14. Either use the result of Problem (7.11.13) or do the calculations directly thus. We have that JP>(Mn :=oxn/rr)
= {~ +~tan- 1
c:) r=
{1-
~tan- 1 (;J
r
if x > 0, by elementary trigonometry. Now tan- 1 y = y + o(y) as y--+ 0, and therefore as n --+ oo.
339
[7.11.15]-[7.11.16]
Solutions
Convergence of random variables
15. The characteristic function of the average satisfies as n--+ oo. By the continuity theorem, the average converges in distribution to the constant JL, and hence in probability also.
16. (a) With Un = u(xn), we have that
n
n
if /lu/1 00 :::0 1. There is equality if Un equals the sign of fn - gn. The second equality holds as in Problem (2.7.13) and Exercise (4.12.3). (b) Similarly, if /lu/1 00 :::0 I, IE(u(X))- E(u(Y))I :::0
L:
lu(x)l · lf(x)- g(x)l dx :::0
L:
lf(x)- g(x)l dx
with equality if u(x) is the sign of f(x)- g(x). Secondly, we have that llP'(X
A) -lP'(Y
E
E
A) I
=
i IE(u(X))- E(u(Y))I :::0 idTV(X, Y),
where u(x)- {
1
if X E A,
-1
ifx¢.A.
Equality holds when A= {x E ffi(: f(x) :::: g(x)}. (c) Suppose dTv(Xn, X)--+ 0. Fix a E ffi(, and let u be the indicator function of the interval ( -oo, a]. Then IE(u(Xn))- E(u(X))I = llP'(Xn :::0 a) -lP'(X :::0 a)l, and the claim follows. On the other hand, if Xn = n- 1 with probability one, then Xn ..!?.. 0. However, by part (a), dTV(Xn, 0) = 2 for all n. (d) This is tricky without a knowledge of Radon-Nikodym derivatives, and we therefore restrict ourselves to the case when X and Y are discrete. (The continuous case is analogous.) As in the solution to Exercise (4.12.4), lP'(X =I Y) ::=:: idTv(X, Y). That equality is possible was proved for Exercise (4.12.5), and we rephrase that solution here. Let J.Ln = min{fn, gn} and JL = Ln J.Ln, and note that dTV(X, Y) = L lfn- gnl = LUn + gn- 2J.Ln} = 2(1- JL). n
n
It is easy to see that 1
z:dTV(X, Y) = lP'(X
=f
Y) =
{
1 if JL = 0, ifJL=l,
O
and therefore we may assume that 0 < JL < I. Let U, V, W be random variables with mass functions lP'(u _ -
Xn
) _ J.Ln JL
lP'(V _ '
-
Xn
) _ max{fn- gn, 0} lP'(W _ ' l-JL
) _ - min{fn- gn, 0} Xn
-
1-J.L
'
and let Z be a Bernoulli variable with parameter JL, independent of (U, V, W). We now choose the pair X', Y' by (X' Y') = { (U, U) if Z = I, ' (V,W) ifZ=O.
340
Solutions [7.11.17]-[7.11.19]
Problems
It may be checked that X' andY' have the same distributions as X andY, and furthermore, li"(X' Y 1) = li"(Z = 0) = 1- fi = idTV(X, Y).
i=
(e) By part (d), we may find independent pairs (X;, Y[), 1 :::0 i :::0 n, having the same marginals as (Xi, Yi), respectively, and such thatli"(X;
:::: 211" (
i=
Y[) = idTV(Xi, Yi). Now,
n
n
L x; i= L:rf i=l
)
n ::::2 Lll"(x;
i=l
i= Y[)
i=l
n
= 2
LdTV(Xj, Yj). i=l
17. If XJ, X2, ... are independent variables having the Poisson distribution with parameter A, then Sn = X1 +X2 + · · · +Xn has the Poisson distribution with parameter An. Now n- 1sn -SA, so that IE(g(n- 1Sn)) -+ g(A) for all bounded continuous g. The result follows.
18. The characteristic function 1/Jmn of (Xn - n)- (Ym - m) ~
Umn=
vm+n
satisfies log 1/lmn (t) = n (eit!.../m+n -
1)
+ m (e-it!.../m+n -
1)
+ (m - n)it -+ _lt2 2 Jm+n
as m, n-+ oo, implying by the continuity theorem that Umn -S N(O, 1). Now Xn + Ym is Poissondistributed with parameter m + n, and therefore
_ v:mn-
J+
Xn Ym P 1 -+ m+n
asm,n-+ oo D
by the law oflarge numbers and Problem (3). It follows by Slutsky's theorem (7.2.5a) that Umn IVmn -+ N(O, 1) as required.
19. (a) The characteristic function of Xn is 1/Jn (t) = exp{i f-tnt - ia;t 2 } where fin and a; are I 2
the mean and variance of Xn. Now, limn-+oo 1/Jn(l) exists. However 1/Jn(l) has modulus e-'Ian, and therefore a 2 = limn-+oo a; exists. The remaining component eitLnt of 1/Jn (t) converges as n -+ oo, say eitLnt -+ 8(t) as n -+ oo where (t) lies on the unit circle of the complex plane. Now
e
I 2 2
1/Jn(t) -+ 8(t)e-'Ia which is required to be a characteristic function; therefore e is a continuous function oft. Of the various ways of showing that 8(t) = eitLt for some f-t, here is one. The sequence 1/Jn(t) = eitLnt is a sequence of characteristic functions whose limit 8(t) is continuous at t = 0. Therefore e is a characteristic function. However 1/1n is the characteristic function of the constant fin, which must converge in distribution as n -+ oo; it follows that the real sequence {/In} converges to some limit f-t, and 8(t) = eitLt as required. This proves that 1/Jn(t)-+ exp{ifl,t- ia 2 t 2 }, and therefore the limit X is N(fl,, a 2). 1 ,
(b) Each linear combinations Xn + t Yn converges in probability, and hence in distribution, to s X +t Y. Now s Xn + t Yn has a normal distribution, implying by part (a) that s X+ t Y is normal. Therefore the joint characteristic function of X and Y satisfies
1/Jx,r(s, t) = I/JsX+tr(1) = exp{iE(sX + tY)- i var(sX + tY)}
ap)}
= exp{i(sf-tx + t{ty)- i
341
[7.11.20]-[7.11.21]
Solutions
Convergence of random variables
in the natural notation. Viewed as a function of s and t, this is the joint characteristic function of a bivariate normal distribution. When working in such a context, the technique of using linear combinations of Xn and Yn is sometimes called the 'Cramer-Wold device'.
20. (i) Write n --+ oo,
=X; -E(X;) and
Y;
2
2
E(Tn fn )
Tn
It suffices to show that n- 1Tn ~ 0. Now, as
I:t=l Y;.
=
I n 2 = 2 l:var(X;) + 2 n i=1 n
(ii) Let E > 0. There exists I such that lp(X;,
2:::
i,J=l
nc
2--+ 0.
Xj) ::0
n
l::Oi<} ::on
Xj )I
::: E if li - jl
n
2::: cov(X;, X1)::: 2:::
cov(X;,
2:::
cov(X;,Xj)+
li-}19 l::Oi,J::On
::=:: /.
Now
cov(X;,X1):::2n/c+n 2 Ec,
li-}1>1 1::oi,}::On
sincecov(X;, Xj)::: lp(Xi, Xj)lyivar(X;) · var(Xj)· Therefore, 2
2/c n
2
E(Tn / n ) ::0 -
+ EC --+ EC
as n--+ oo.
This is valid for all positive E, and the result follows.
21. The integral r)O_c_dx 12 x log lxl diverges, and therefore IE( X I) does not exist. The characteristic function cp of X 1 may be expressed as ,~., )
'f'(t
= 2C
100 2
whence cp(t)- cp(O) = _ [ 2c 12
cos(tx) d X x 2 Iogx
---
00 1 -
cos(tx) dx. x 2 Iogx
Now 0::: I- cose::: min{2, 8 2 }, and therefore
Icp(t)-2c cp(O) I -< Now
11/t 2
11u
-
u
and
2
!00
t2 2 --dx+ ---dx Iogx 1/t x2Iogx '
dx ----+0
Therefore
as u--+ oo,
Iogx
2 1oo 2 dx - - dx 1uoo -x2Iogx -Iogu u x 2 I < --
if t > 0.
-
Icp(t) ~ cp(O) I = o(t)
2 = --
ulogu'
as t
t
u > I.
0.
Now cp is an even function, and hence cp' (0) exists and equals 0. Use the result of Problem (7.ll.l5) to deduce that n- 1 I:'i Xi converges in distribution to 0, and therefore in probability also, since 0 is constant. The X; do not obey the strong law since they have no mean.
342
Solutions [7.11.22]-[7.11.24]
Problems
22. If the two points are U and V then
and therefore 12 1~ 2p1 -Xn = - L)Ui -Vi) -+ n n i= 1 6
as n
oo,
--+
by the independence of the components. It follows that Xn/ ,fii ~ Problem (7 .11.3) or by the fact that
1/ .J6 either by the result of
23. The characteristic function of Yj = Xj 1 is 4J(t) = 1
. + e-ztfx) . dx = lo1 cos(tjx) dx =It I100 -cosy- dy lo 1(eztfx ltl Y2
2 0
0
by the substitution x = It II y. Therefore l/J(t) = 1- It I
100 ltl
1- cosy
y
2
as t--+ 0,
dy = 1- /It I+ o(ltl)
where, integrating by parts,
-looo 1 - cos y dy--looo -sin-u du-_ -.
I-
11:
0
y
2
0
u
2
It follows that Tn = n- 1 '£,)= 1 Xj 1 has characteristic function
as t--+ oo, whence 2Tn/7r is asymptotically Cauchy-distributed. In particular, 1P'(I2Tn/rrl >
2100
1)--+ -
11:
1
du 1+u
1 2
as t
--2 =-
--+
oo.
24. Let mn be a non-decreasing sequence of integers satisfying 1 ::: mn < n, mn
noting that Ynk takes the value ±1 each with probability '£Z= 1 Ynk· Then n
lP'(Un-:/= Zn) :"':
LlP'(IXkl ~
i whenever mn
--+
< k < n. Let Zn
n
mn) :"':
L k=mn
k=1
343
k2
--+
0
oo, and define
as n--+ oo,
[7.11.25]-[7.11.26]
Solutions
Convergence of random variables
from V.:hich it follows that Un/ Jn
-S N(O, 1) if and only if Zn/ Jn -S N(O, 1). Now mn
L
Zn =
Ynk
+ Bn-mn
k=1
where Bn-mn is the sum of n- mn independent summands each of which takes the values ±1, each possibility having probability Furthermore
1·
I
1
1 mn 2 -2:Ynk ::S mn Jn k=1 Jn
which tends to 0 if mn is chosen to be mn = n- 1 Bn-mn
Ln 1/ 5J; with this choice for mn, we have that
-S N(O, 1), and the result follows.
Finally, var(Un)
=
t (2 - ~)
k=1
k
so that 1 n 1 var(Un/ Jn) = 2- 2 --+ 2. n k=1 k
L
25. (i) Let c/Jn and cp be the characteristic functions of Xn and X. The characteristic function !frk of XNk is 00
!frk(t)
= 2::: cfJ1 (t)f'(Nk = J) }=1
whence 00
11/fk(t)-
cp(t)l:::::
L lc/Jj(t)- cp(t)llP'(Nk = j). }=1
Let E > 0. We have that ifJ} (t) --+ cp (t) as j --+ oo, and hence for any T > 0, there exists J (T) such that lifJ} (t) - cp(t)l < E if j :::: J (T) and It I ::::: T. Finally, there exists K(T) such that lP'(Nk ::::: J (T)) ::::: E if k :::: K (T). It follows that
if It I ::::: T and k :::: K(T); therefore !frk(t) --+ cp(t) ask--+ oo. (ii) Let Yn = SUPm;:::n IXm -XI. ForE > 0, n :0:: 1, lP'(IXNk- XI>
E)::::: lP'(Nk::::: n) + lP'(IXNk- XI> E, Nk ::::: lP'(Nk ::::: n)
+ lP'(Yn
> n)
> E) --+ lP'(Yn > E)
Now take the limit as n --+ oo and use the fact that Yn ~ 0.
26. (a) We have that ( a(n
k
)
an- -,n - =
+ 1, n)
ITk ( 1 i=O
"'l
k .) -1. ) < exp ( - L..J - . n i=O n
344
ask --+
00.
Solutions
Problems
e
(b) The expectation is
L_g
En=
Jnn)
J
where the sum is over all j satisfying n - M Jn
(j-
g
:::
(nj+l _
Jn
j!
nl;~n
j ::: n. For such a value of j,
n) nle-n =e-n
Jn
[7.11.27]-[7.11.28]
nl ) (j- l)! ,
j!
whence En has the form given. (c) Now g is continuous on the interval [ -M, 0], and it follows by the central limit theorem that
1
I I 2 g(x)--e-'1.x dx
0
En--+
.J2ii
- M
I
x I 2 --e-'1.x dx
!oM
=
=
.J2ii
0
1-e-'J.M
2
------.,=-
.J2ii
Also,
where k
= LM JnJ.
Take the limits as n --+ oo and M --+ oo in that order to obtain 1
.
--
.J2ii -
nn+2e-n I
n--+ oo {
n
1
}
< --.
!
-
.J2ii
27. Clearly E(Rn+1 I Ro, R1, ... , Rn)
+ 2).
since a red ball is added with probability Rn/(n
Rn n+2
= Rn + - -
Hence
and also 0 ::: Sn ::: 1. Using the martingale convergence theorem, S = limn--+oo Sn exists almost surely and in mean square.
28. Let 0 <
E
< ~, and let
k(t)
=
L8tJ, m(t)
= ro- E3)k(t)l,
n(t)
=
L(l
+ E3)k(t)J
and let Imn(t) be the indicator function of the event {m(t) ::: M(t) < n(t)}. Since M(t)jt ~ e, we may find T such that EUmn(t)) > I-E fort~ T. We may approximate SM(t) by the random variable sk(t) as follows. With Aj = {ISj - sk(t) I >
E.Jk(t)}, IP'(AM(t)) ::0 IP'(AM(t)• lmn(t) k(t)-1
= 1) + IP'(AM(t)• lmn(t) = 0) n(t)-1
:::lP'( U Aj)+lP'( U j=m(t) < -
Aj)+IP'(lmn(t)=O)
j=k(t)
{k(t) - m(t) }a 2 E2k(t)
+ if t
345
{ n(t)- k(t) }a 2
E2k(t) ~
T,
+E
[7.11.29]-[7.11.32]
Solutions
Convergence of random variables
by Kolmogorov's inequality (Exercise (7.8.1) and Problem (7.11.29)). Send t --+ oo to find that as t --+ oo.
Now Sk(t)/ .Jk(i)
~
N(O, a 2 ) as t --+ oo, by the usual central limit theorem. Therefore
which implies the first claim, since k(t)j(et) --+ 1 (see Exercise (7.2.7)). The second part follows by Slutsky's theorem (7.2.5a).
29. We have that Sn = Sk
+ (Sn -
Sk), and so, for n :::: k,
Now S'fJAk :::: c 2 IAk; the second term on the right side is 0, by the independence of the X's, and the third term is non-negative. The first inequality of the question follows. Summing over k, we obtain E(S;) :::: c 2 1P'(Mn > c) as required.
30. (i) With Sn = Lt=1 xj, we have by Kolmogorov's inequality that 1 m+n lP' ( max ISm+k- Sml > E) ::0 2 E(X~) 1~k~n E k=m
L
forE > 0. Take the limit as m, n --+ oo to obtain in the usual way that {Sr : r :::: 0} is a.s. Cauchy convergent, and therefore a.s. convergent, if 2::\"' E(X~) < oo. It is shorter to use the martingale convergence theorem, noting that Sn is a martingale with uniformly bounded second moments. (ii) Apply part (i) to the sequence Yk = Xkfbk to deduce that 2::~ 1 Xkfbk converges a.s. The claim now follows by Kronecker's lemma (see Exercise (7.8.2)).
31. (a) This is immediate by the observation that ei.(P) =
f Xo IJ PN_ij lJ · i,j
I
(b) Clearly Lj Pij = 1 for each i, and we introduce Lagrange multipliers {JL; : i E S} and write V = A.(P) + Li JLi Lj Pij. Differentiating V with respect to each Pij yields a stationary (maximum) value when (N;j / Pij) + JLi = 0. Hence Lk N;k = - JLb and
z=E\
(c) We have that Nij = Nik Ir where Iris the indicator function oftheeventthatthe rth transition out ofi is to j. By the Markov property, the Ir are independent with constant mean Pij. Using the strong law oflarge numbers and the fact that Lk N;k ~ oo as n --+ oo, Pij ~ E(/t) = Pij. 32. (a) If X is transient then V;(n) < oo a.s., and JLi = oo, whence V;(n)/n ~ 0 = ttj 1. If X is persistent, then without loss of generality we may assume Xo = i. LetT (r) be the duration of the rth
346
Solutions [7.11.33]-[7.11.34]
Problems
excursion from i. By the strong Markov property, the T (r) are independent and identically distributed with mean J.Li . Furthermore, V;(n)-J
::r
V;(n)
-
n
1
V;(n)
::r ·
T(r) < < " V;(n) - V;(n)
"
T(r)
By the strong law of large numbers and the fact that V; (n) ~ oo as n --+ oo, the two outer terms sandwich the central term, and the result follows. (b) Note that ~~,;;;J f(Xr) = ~iES f(i)V;(n). With Q a finite subset of S, and n; = J.Lj 1, the unique stationary distribution,
::":
{"I
V;(n) 1 L.t ------:-
iEQ
n
J.Lz
I
1 )} 11/lloo, + "(V;(n) L.t - - +---:i
n
J.Lz
where 11/lloo = sup{l/(i)l : i E S}. The sum over i E Q converges a.s. to 0 as n--+ oo, by part (a). The other sum satisfies "(V;(n) L.t - i
1) -_2- L.t
"(Vi(n) -iEQ n
+-
J.Li
which approaches 0 a.s., in the limits as n --+ oo and Q
t
+ lrj )
S.
33. (a) Since the chain is persistent, we may assume without loss of generality that Xo = j. Define the times RJ, R2, ... of return to j, the sojourn lengths SJ, S2, ... in j, and the times VJ, V2, ... between visits to j. By the Markov property and the strong law of large numbers, 1
n "S
-~
a.s.
1 1~ a.s. - Rn = - L.t Vr ----+ J.Lj. n n r=l
1
r~-,
n r=l
gj
Also, Rn/ Rn+J ~ 1, since J.Lj = E(RJ) < oo. If Rn < t < Rn+l• then
Let n --+ oo to obtain the result. (b) Note by Theorem (6.9.21) that Pij (t) --+ nj as t --+ oo. We take expectations of the integral in part (a), and the claim follows as in Corollary (6.4.22). (c) Use the fact that
lot f(X(s))ds L l =
0
l(X(s)=jJdS
jES 0
together with the method of solution of Problem (7 .11.32b).
34. (a) By the first Borel-Cantelli lemma, Xn = Yn for all but finitely many values of n, almost surely. Off an event of probability zero, the sequences are identical for all large n. (b) This follows immediately from part (a), since Xn - Yn = 0 for all large n, almost surely. (c) By the above, a;;- 1 ~~~ (Xr - Yr) ~ 0, which implies the claim.
347
[7.11.35]-[7.11.37]
Solutions
Convergence of random variables
35. Let Yn = Xnl(IXn[.:Sa)· Then,
n
n
by assumption (a), whence {Xn} and {Yn} are tail-equivalent (see Problem (7.11.34)). By assumption (b) and the martingale convergence theorem (7.8.1) applied to the partial sums 2:::~= 1 (Yn - IE(Yn)), the infinite sum 2:::~ 1 (Yn - IE(Yn)) converges almost surely. Finally, I:~ 1 IE(Yn) converges by assumption (c), and therefore 2:::~ 1 Yn, and hence 2:::~ 1 Xn, converges a.s.
36. (a) Let n1 < n2 < · · · < nr = n. Since the
h take only two values, it suffices to show that r
lP'(/ns
= 1 for 1 .::5 s
.::5 r)
=
IT lP'Uns = 1). s=1
Since F is continuous, the Xi take distinct values with probability 1, and furthermore the ranking of X 1, X2, ... , Xn is equally likely to be any of the n! available. Let x1, x2, ... , Xn be distinct reals, and write A= {Xi =Xi for 1 .::5 i .::5 n}. Now, lP'Uns = 1 for 1 .::5 s .::5 r
I A)
1{(nns-1 - 1) (n- 1- ns-1)!· }{(ns-ns-21 - 1) (ns-1- 1- ns-2)! }· · · (n1- 1)!
=I n.
1
1
1
and the claim follows on averaging over the Xi. (b)WehavethatlE(h) = lP'(h = 1) = k- 1 and var(h) = r 1(1-k- 1), whence l::k var(h/logk) < oo. By the independence of the h and the martingale convergence theorem (7.8.1), 2:::~ 1 (h k- 1) flog k converges a.s. Therefore, by Kronecker's lemma (see Exercise (7.8.2)),
t(/·-~) ~0 1 J
- 1 logn }= 1 The result follows on recalling that
as n---+ oo.
I:,}= 1 j - 1 ~ log n as n ---+
oo.
37. By an application of the three series theorem of Problem (7.11.35), the series converges almost surely.
348
8
Random processes
8.2 Solutions. Stationary processes 1.
With ai (n)
= lP'(Xn = i), we have that
cov(Xm, Xm+n) = lP'(Xm+n = 1 I Xm = 1)1P'(Xm = 1) -lP'(Xm+n = 1)1P'(Xm = 1)
= a1 (m)pu (n)- al (m)al (m + n), and therefore, a1 (m)pu (n)- a1 (m)al (m
+ n)
p(Xm,Xm+n)=-r=r~~~~~~~~~==T=~~
v'a1 (m)(l - a1 (m))al (m
Now, a1 (m)
~
+ n)(l -
a1 (m
+ n))
aj(a + fJ) as m ~ oo, and a a+fJ
Pll (n) = - whence p(Xm, Xm+n)
~
+ -fJ- (1 a+fJ
n
-a - fJ) ,
(1 -a- fJ)n as m ~ oo. Finally,
. 1 n hm - LlP'(Xr
=
a
1)
= --. a+ fJ
n-+oo n r=l
The process is strictly stationary if and only if Xo has the stationary distribution.
2.
We have that E(T(t)) = 0 and var(T(t)) = var(To) = 1. Hence: (a) p(T(s), T(s + t)) = E(T(s)T(s + t)) = E[(-1)N(t+s)-N(sl] = e-2M.
(b) Evidently, E(X (t)) = 0, and E[X(t) 2 ] = E ( l l T(u)T(v)dudv)
E(T(u)T(v))dudv=21 1 1v e- 2}..(v-u)dudv ~
=2 { =
3.
I1 ( t -
1 2A.
+ 21A. e -2At)
We show first the existence of the limitA. g(x
+ y) =
~
It
as t ~ oo.
= limq.o g(t)jt, where g(t) = lP'(N(t)
+ y) > 0) = lP'(N(x) > 0) + lP'({N(x) = 0} n for x, y 2: 0. :S g(x) + g(y)
> 0). Clearly,
lP'(N(x
349
{N(x
+ y)- N(x)
> 0})
[8.3.1]-[8.3.4]
Solutions
Random processes
Such a function g is called subadditive, and the existence of A follows by the subadditive limit theorem discussed in Problem (6.15.14). Note that A= oo is a possibility. Next, we partition the interval (0, 1] into n equal sub-intervals, and let ln(r) be the indicator function of the event that at least one arrival lies in ( (r - 1) In, r In] , 1 .::5 r .::5 n. Then 2:::~= 1 In (r) t N(l) as n ---+ oo, with probability 1. By stationarity and monotone convergence, E(N(l)) =
m:( lim
~ In(r))
n-+oo L....t
=
lim
n-+oo
r=I
m:(~ In(r)) L....t
= lim ng(n- 1) =A. n-+oo
r=I
8.3 Solutions. Renewal processes 1.
See Problem (6.15.8).
2.
With X a certain inter-event time, independent of the chain so far,
Bn+I = {
X - 1 if Bn = 0, Bn - 1 if Bn > 0.
Therefore, B is a Markov chain with transition probabilities Pi,i-I = 1 fori > 0, and POJ = fJ+I for j 2: 0, where fn = lP'(X = n). The stationary distribution satisfies TCJ = TCJ+J + nof}+J, j 2: 0, with solution n1 = lP'(X > j)IE(X), provided E(X) is finite. The transition probabilities of B when reversed in equilibrium are
Pi,i+I
TCi+I
= 7; =
lP'(X > i + 1) lP'(X > i)
fi+I PiO = lP'(X > i),
fori 2: 0.
These are the transition probabilities of the chain U of Exercise (8.3.1) with the Jj as given.
3. We have that pnun = 2:::~=! pn-kUn-kPk fk, whence Vn = pnun defines a renewal sequence provided p > 0 and l:n pn fn = 1. By Exercise (8.3.1 ), there exists a Markov chain U and a state s such that Vn = lP'(Un = s)---+ ns, as n---+ oo, as required.
4.
Noting that N(O) = 0, oo
oo
oo
oo
k=I
r=k
r
L E(N(r))sr = L L UkSr = L Uk L sr r=O
r=I k=I
= ~ uvk = U(s)- 1 = F(s)U(s).
L....t1-s k=I
1-s
1-s
Let Sm = l:k=I Xk and So= 0. Then lP'(N(r) = n) = lP'(Sn .::5 r) -lP'(Sn+I .::5 r), and
~sfE 00
[(N(t)k
+k)] =~sf~ n +k) k (lP'(Sn .::5 t) -lP'(Sn+I .::5 t)) 00
00
(
= ~sf 00
[
1
+
E
Now,
350
00
(
n+ k _k
~
1)
]
lP'(Sn .::5 t) .
Solutions
Queues
[8.3.5]-[8.4.4]
whence, by the negative binomial theorem, "'too=OstlE
L...J
[(N(t)k+k)] =
1
(1- s)(l - F(s))k
U(d 1-s
5. This is an immediate consequence of the fact that the interarrival times of a Poisson process are exponentially distributed, since this specifies the distribution of the process.
8.4 Solutions. Queues 1. We use the lack-of-memory property repeatedly, together with the fact that, if X and Y are independent exponential variables with respective parameters 'A and J-t, then JP>(X < Y) = 'A/('A + J-t). (a) In this case, 1{
p
=2
'A 1-t 1-t } 1 { 1-t 'A 'A } 'A + J-t . 'A + J-t + 'A + J-t + 2 'A + J-t . 'A + J-t + 'A + J-t
(b) If 'A< f.-t, and you pick the quicker server, p
= 1-
(-1-t-)
1
= 2+
2'A~-t ('A +
~-t)2.
2
"A+~-t
2'A~-t
(c) And finally, p = ('A+ J-t) 2 .
2. The given event occurs if the time X to the next arrival is less than t, and also less than the time Y of service of the customer present. Now,
3.
By conditioning on the time of passage of the first vehicle, IE(T) =loa (x
+ IE(T) )'Ae-1.x dx + ae-1.a,
and the result follows. If it takes a time b to cross the other lane, and so a + b to cross both, then, with an obvious notation, (a) (b)
eal.- 1 ebJ.L- 1 IE(Ta) +IE(n) = - - + - - , 'A 1-t e
The latter must be the greater, by a consideration of the problem, or by turgid calculation.
4.
Look for a solution of the detailed balance equations f.-tlTn+l
=
'A(n + 1) n + 2 Jl'n,
n:::: 0.
tofindthatnn = pnno/(n+1)isastationarydistributionifp < l,inwhichcaseno = -pjlog(l-p). Hence l:n nnn = 'Ano/(J-t- 'A), and by the lack-of-memory property the mean time spent waiting for service is pno/(J-t- 'A). An arriving customer joins the queue with probability
L
00
n + 1lTn = p + log(l- p). n=O n + 2 p log(l - p)
351
[8.4.5]-[8.5.5]
Solutions
Random processes
5. By considering possible transitions during the interval (t, t +h), the probability Pi (t) that exactly i demonstrators are busy at time t satisfies: P2(t +h)= P1 (t)2h
+ P2(t)(l
- 2h)
+ o(h),
P1 (t +h)= Po(t)2h + P1 (t)(l - h)(l - 2h) + P2(t)2h + o(h), Po(t +h)= Po(t)(l- 2h)
+ P1(t)h + o(h).
Hence, P~ (t)
= 2pt (t)
- 2p2(t),
Pi (t)
= 2po(t) -
Jp1 (t)
+ 2p2 (t),
0
p (t)
=
-2po(t)
+ Pt (t),
and therefore P2(t) =a+ be- 2t + ce-St for some constants a, b, c. By considering the values of P2 and its derivatives at t = 0, the boundary conditions are found to be a + b + c = 0, - 2b - 5c = 0, 4b + 25c = 4, and the claim follows.
8.5 Solutions. The Wiener process 1. We might as well assume that W is standard, in that a 2 = 1. Because the joint distribution is multivariate normal, we may use Exercise (4.7.5) for the first part, and Exercise (4.9.8) for the second, giving the answer -1 + 1 { sm. 1 8 4n 2.
=
Writing W(s)
..fiX, W(t)
If
-+sin- 1 t
= ,fiz,
If
and W(u)
-+sin- 1 u
=
If} u
.
.JUY, we obtain random variables X, = ..[ilU, P2 = ..filU,
Y, Z with the standard trivariate normal distribution, with correlations P1
P3 = .JS[i. By the solution to Exercise (4.9.9),
var(Z I X, Y) = yielding var(W(t)
I W(s),
(u-t)(t-s) t(u- s)
,
W(u)) as required. Also,
I
E(W(t)W(u) W(s), W(v)) = E {[
I
(u-t)W(s)+(t-s)W(u)] u_ s W(u) W(s), W(v)
},
which yields the conditional correlation after some algebra. 3.
Whenever a 2 + b 2 = 1.
4.
Let lij(n)
= W((j +
1)tfn)- W(jtfn). By the independence of these increments,
5. They all have mean zero and variance t, but only (a) has independent normally distributed increments.
352
Solutions [8.7.1]-[8.7.3]
Problems
8.7 Solutions to problems 1. E(Yn) = 0, and cov(Ym, Ym+n) = Ll=O CXiCXn+i form, n 2': 0, with the convention that CXk fork> r. The covariance does not depend on m, and therefore the sequence is stationary.
=0
2. We have, by iteration, that Yn = Sn (m) +am+ 1 Yn-m-1 where Sn (m) = LJ=O cxi Zn- j. There are various ways of showing that the sequence {Sn (m) : m ::: 1} converges in mean square and almost surely, and the shortest is as follows. We have that cxm+ 1Yn-m-1 ---+ 0 in m.s. and a.s. as m ---+ oo; to see this, use the facts that var(cxm+ 1 Yn-m-1) = cx 2(m+ 1) var(Yo), and
E
> 0.
It follows that Sn(m) = Yn- am+ 1Yn-m-1 converges in m.s. and a.s. as m---+ oo. A longer route to the same conclusion is as follows. For r < s,
whence {Sn(m) : m ::: 1} is Cauchy convergent in mean square, and therefore converges in mean square. In order to show the almost sure convergence of Sn(m), one may argue as follows. Certainly
whence Z:::f:=o cxi Zn- j is a.s. absolutely convergent, and therefore a.s. convergent also. We may express limm-+oo Sn (m) as Z:::f:=o cxi Zn- j. Also, am+ 1Yn-m-1 ---+ 0 in mean square and a.s. as m ---+ oo, and we may therefore express Yn as 00
Yn =
L cxi Zn-
j
a.s.
j=O
It follows that E(Yn) = limm-+oo E(Sn(m)) = 0. Finally, for r > 0, the autocovariance function is given by c(r) = cov(Yn, Yn-r) = E{ (cxYn-1 + Zn)Yn-r} = cxc(r - 1),
whence cxlrl
c(r)
= cxlrlc(O) = - -2 , l-ex
r= ... ,-1,0,1, ... ,
since c(O) = var(Yn). 3. N(t)
1ft is a non-negative integer, N(t) is the number ofO's and 1's preceding the (t+ l)th 1. Therefore
+ 1 has the negative binomial distribution with mass function
k:::t+l.
If tis not an integer, then N(t) = N(Lt J).
353
[8.7.4]-[8.7.6] 4.
Solutions
Random processes
We have that A.h
JP>(Q(t +h)= j I Q(t)
= i) = {
+ o(h) + o(h)
+ 1, = i - 1,
if j = i if j
11-ih
1 -(A.+ 11-i)h
+ o(h)
if j = i,
an immigration-death process with constant birth rate A. and death rates /1-i = i fl-. Either calculate the stationary distribution in the usual way, or use the fact that birth-death processes are reversible in equilibrium. Hence 'Ani = 11-U + l)rri+l fori 2:: 0, whence
n·l = -.,1 I.
(A.)i e-A/JL 11-
i :::: 0.
'
5. We have that X(t) = R cos(IJ!) cos(&t) - R sin(IJ!) sin(&t). Consider the transformation u r cos 1/f, v = -r sin 1/f, which maps [0, oo) x [0, 2n) to R 2 . The Jacobian is au
au
ar
a!fr
ar
aljf
av
av
=
= -r,
whence U = R cos IJf, V = - R sin IJf have joint density function satisfying rfu,v(rcosljf, -rsin 1/f)
Substitute fu,v(u, v)
= fR,\jl(r, 1/f).
= e-2I (u 2+v 2) /(2rr), to obtain r > 0, 0::::: 1/f < 2n.
Thus Rand IJf are independent, the latter being uniform on [0, 2rr). 6. A customer arriving at time u is designated green if he is in state A at time t, an event having probability p(u, t - u). By the colouring theorem (6.13.14), the arrival times of green customers form a non-homogeneous Poisson process with intensity function A.(u)p(u, t - u), and the claim follows.
354
9
Stationary processes
9.1 Solutions. Introduction 1.
We examine sequences Wn of the form 00
Wn
=
LakZn-k k=O
for the real sequence {ak : k 2:: 0}. Substitute, to obtain ao = l, a1 =a, ar = aar-1 with solution
+ fJar-2• r
2:: 2,
otherwise, where A. 1 and A.2 are the (possibly complex) roots of the quadratic x 2 -ax - fJ = 0 (these roots are distinct if and only if a 2 + 4{3 =!= 0). Using the method in the solution to Problem (8.7.2), the sum in(*) converges in mean square and almost surely if IA.1I < 1 and IA.2I < 1. Assuming this holds, we have from(*) that JE(Wn) = 0 and the autocovariance function is c(m) = JE(Wn Wn-m) = ac(m- 1)
+ {Jc(m- 2),
m
::=::
1,
by the independence of the Zn. Therefore W is weakly stationary, and the autocovariance function may be expressed in terms of a and fJ.
2.
We adopt the convention that, if the binary expansion of U is non-unique, then we take the (unique) non-terminating such expansion. It is clear that Xi takes values in {0, 1}, and
for all x1, x2, ... , Xn; therefore the X's are independent Bernoulli random variables. For any sequence k1 < k2 < ... < kr, the joint distribution of vk,' vk2' ... ' vkr depends only on that of Xk 1+1, Xk 1+2• .... Since this distribution is the same as the distribution of X 1, X2, ... , we have that (Vk 1 , Vk2 , ... , Vkr) has the same distribution as (Vo, Vk2-k1 , •.• , Vkr-k! ). Therefore V is strongly stationary. Clearly JE(Vn) = JE(Vo) = i. and, by the independence oftlie Xi, 00
cov(Vo, Vn)
= 2:T 2i-n var(Xi) = i=1
355
tz(i)n.
[9.1.3]-[9.2.1]
Solutions
Stationary processes
3. (i) For mean-square convergence, we show that Sk = L~=O anXn is mean-square Cauchy convergent as k -+ oo. We have that, for r < s, lE{ (Ss - Sr ) 2 }
= . _L s
{
L s
aiajc(i - j) .::; c(O) . lai I z,;=r+I l=r+I
}2
since lc(m) I .::; c(O) for all m, by the Cauchy-Schwarz inequality. The last sum tends to 0 as r, s -+ oo if Li lai I < oo. Hence Sk converges in mean square ask -+ oo. Secondly, lE(
t
lakXkl) .::::
k=I
t
lakl ·lEIXkl.:::: VJE
k=I
t
lakl
k=I
which converges as n -+ oo if the lakl are summable. It follows that Lk=I lakXkl converges absolutely (almost surely), and hence Lk=I akXk converges a.s. (ii) Each sum converges a.s. and in mean square, by part (i). Now 00
cy(m) =
L
ajakc(m + k- j)
j,k=O whence
L lcy(m)l.:::: c(O){ f
2
laj I} < oo.
j=O
m
4. Clearly Xn has distribution :rr for all n, so that {f(Xn) : n :::: m} has fdds which do not depend on the value of m. Therefore the sequence is strongly stationary.
9.2 Solutions. Linear prediction 1.
(i) We have that
which is minimized by setting a = c(l)/c(O). Hence Xn+I = c(l)Xn/c(O). (ii) Similarly
an expression which is minimized by the choice
f3
c(l) (c(O)- c(2)) = c(0) 2 - c(1)2 '
c(O)c(2) - c(1) 2 y = c(0)2 - c(1)2 ;
Xn+I is given accordingly. (iii) Substitute a, {3, y into(*) and(**), and subtract to obtain, after some manipulation, {c(l ) 2 - c(O)c(2) } 2 D - .:-,--:-~..-'--'--.:...:.:.: - c(O){c(0)2- c(l)2} ·
356
Solutions
Autocovariances and spectra 1
(a) In this case c(O) = '2> and c(l) (b) In this caseD = 0 also.
= c(2) = 0.
-
Therefore Xn+1
[9.2.2]-[9.3.2]
= Xn+1 = 0, and D = 0.
In both (a) and (b), little of substance is gained by using Xn+1 in place of Xn+1· 2. Let {Zn : n = ... , -1, 0, 1, ... } be independent random variables with zero means and unit variances, and define the moving-average process
Zn + aZn-1 ~.
Xn=
It is easily checked that X has the required autocovariance function.
By the projection theorem, Xn - Xn is orthogonal to the collection {Xn-r : r > 1}, so that 00 . E{(Xn- Xn)Xn-r} = 0, r :=:: 1. Set Xn = l::s=1 hsXn-s to obtam that for s:::: 2, where a = aj(l + a 2 ). The unique bounded solution to the above difference equation is hs (-l)s+ 1as, and therefore 00
Xn
= 2:)-l)s+ 1as Xn-s· s=1
The mean squared error of prediction is
Clearly E(Xn) = 0 and 00
cov(Xn, Xn-m)
=
L
brbsc(m
+ r - s),
m ::::0,
r,s=1
so that
X is weakly stationary. 9.3 Solutions. Autocovariances and spectra
1.
It is clear that E(Xn) = 0 and var(Xn) = 1. Also
cov(Xm, Xm+n) = cos(mA.) cos{(m
+ n)A.} + sin(mA.) sin{(m + n)A.} =
cos(nA.),
so that X is stationary, and the spectrum of X is the singleton {A.}.
2.
Certainly 1/Ju (t)
= (eit:rr
cov(Xm, Xm+n)
- e-it:rr)/(2nit), so that E(Xn)
= E(XmXm+n) =
= 1/Ju (1)1/Jv (n) = 0.
E(ei{U-Vm-U+V(m+n)J)
whence X is stationary. Finally, the autocovariance function is c(n) = 1/Jv(n) =
j
whence F is the spectral distribution function.
357
einJ... dF(A.),
Also
= 1/lv(n),
[9.3.3]-[9.3.4]
Solutions
Stationary processes
3.
The characteristic functions of these distributions are
(i)
p(t)
(ii)
p(t) =
4.
1 2
= e-zt
~
,
c~it
+ 1 ~it)
= 1
~ t2 .
(i) We have that
L
var ( -1 n Xj ) = 21 n }=!
Ln
cov(Xj, Xk) = -c(O) 2
n j,k=1
n
1 (L n
· · ) dF(A.). ez(j-k))..
j,k=1
(-Jr,Jr]
The integrand is 1 - cos(nA.) 1- cos A. ' whence
~X · ) var ( -1 L n }= 1 1 It is easily seen that I sine I .:5
= c(O)
1 ( (-Jr,Jr]
sin(nA./2) ) 2 dF(A.). n sin(A./2)
IB I, and therefore the integrand is no larger than (
A./2 ) 2 1 2 sin(A./2) .:5 (2n) ·
As n --+ oo, the integrand converges to the function which is zero everywhere except at the origin, where (by continuity) we may assign it the value 1. It may be seen, using the dominated convergence theorem, that the integral converges to F (0) - F (0-), the size of the discontinuity of F at the origin, and therefore the variance tends to 0 if and only if F(O) - F(O-) = 0. Using a similar argument,
-1 n-1 L c(j) = n }=0
1
1
L
(0) (n-1 ) _ceij).. dF(A.) = c(O) gn(A.)dF(A.) n (-Jr,Jr] j=O (-Jr,Jr]
where
ifA.
= 0,
ifA.
¥= 0,
gn(A.) = { 1ein)..- 1 n(ei).. - 1)
is a bounded sequence of functions which converges as before to the Kronecker delta function 8;..o. Therefore 1 n-1 c(j) --+ c(O) ( F(O) - F(O-)) as n--+ oo. n }=0
-L
358
Solutions
The ergodic theorem
[9.4.1]-[9.5.2]
9.4 Solutions. Stochastic integration and the spectral representation 1. Let Hx be the space of all linear combinations of the X;, and let H x be the closure of this space, that is, H x together with the limits of all mean-square Cauchy-convergent sequences in H x. All members of Hx have zero mean, and therefore all members of H x also. Now S(A.) E H x for all A., whence JE.(S(A.)- S(fl-)) = 0 for all A. and fl-. 2. First, each Ym lies in the space H x containing all linear combinations of the X n and all limits of mean-square Cauchy-convergent sequences of the same form. As in the solution to Exercise (9 .4.1 ), all members of H x have zero mean, and therefore JE.(Ym) = 0 for all m. Secondly,
_ 1
lE.(YmY n) =
(-n,n]
eimJ...e-inJ... /(A.) dA. = 8mn·
2rrf(A.)
As for the last part,
This proves that such a sequence Xn may be expressed as a moving average of an orthonormal sequence. 3. Let H x be the space of all linear combinations of the Xn, together with all limits of (meansquare) Cauchy-convergent sequences of such combinations. Using the result of Problem (7 .11.19), all elements in H x are normally distributed. In particular, all increments of the spectral process are normal. Similarly, all pairs in H x are jointly normally distributed, and therefore two members of H x are independent if and only if they are uncorrelated. Increments of the spectral process have zero means (by Exercise (9.4.1)) and are orthogonal. Therefore they are uncorrelated, and hence independent.
9.5 Solutions. The ergodic theorem 1. With the usual shift operator r, it is obvious that r- 1 0 = 0, so that 0 E 1. Secondly, if A then r- 1(A c) = (r- 1A)c = Ac, whence Ac E 1. Thirdly, suppose A1, Az, ... E 1. Then
E
1,
so that U'f A; E 1. 2. The left-hand side is the sum of covariances, c(O) appearing n times, and c(i) appearing 2(n- i) times for 0 < i < n, in agreement with the right-hand side. Let E > 0. Ifc(j) = when j 2::: J. Now
r
1 2:{~~
c(i) --+ a 2 as j --+
2n
2{1
n
n
00,
n
2 Ljc(j) :52 LjC(j) + L j=1
j=1
there exists J such that lc(j) - a 2 1 <
j(a2
+E) }
--+ a2
E
+E
j=l+1
as n --+ oo. A related lower bound is proved similarly, and the claim follows since E ( > 0) is arbitrary.
359
[9.5.3]-[9.6.3] 3.
Solutions
Stationary processes
It is easily seen that Sm = l::~o ai Xn+i constitutes a martingale with respect to the X's, and oo
m
E(S~)
=
I>tE(X~+i):::;
L:>f,
i=O
i=O
whence Sm converges a.s. and in mean square as m--+ oo. Since the Xn are independent and identically distributed, the sequence Yn is strongly stationary; also E(Yn) = 0, and so n- 1 l::t=I Yi --+ Z a.s. and in mean, for some random variable Z with mean zero. For any fixed m (2: 1), the contribution of X 1, Xz, ... , Xm towards l::t=l Yi is, for large n, no larger than Cm =
Jf(ao +a,+ ... + aj-I)Xj I· J=l
Now n- 1cm --+ 0 as n --+ oo, so that Z is defined in terms of the subsequence Xm+I, Xm+2, ... for all m, which is to say that Z is a tail function of a sequence of independent random variables. Therefore Z is a.s. constant, and so Z = 0 a.s.
9.6 Solutions. Gaussian processes 1. The quick way is to observe that c is the autocovariance function of a Poisson process with intensity 1. Alternatively, argue as follows. The sum is unchanged by taking complex conjugates, and hence is real. Therefore it equals
where
to= 0.
2. For s, t :::: 0, X (s) and X (s + t) have a bivariate normal distribution with zero means, unit variances, and covariance c(t). It is standard (see Problem (4.14.13)) that E(X(s + t) I X(s)) = c(t)X(s). Now c(s
+ t) = E(X(O)X(s + t)) = E{E(X(O)X(s + t) I X(O), X(s))} = E(X(O)c(t)X(s)) = c(s)c(t)
by the Markov property. Therefore c satisfies c(s + t) = c(s)c(t), c(O) = 1, whence c(s) = c(l)lsl = plsl. Using the inversion formula, the spectral density function is
Note that X has the same autocovariance function as a certain autoregressive process. Indeed, stationary Gaussian Markov processes have such a representation.
3. If X is Gaussian and strongly stationary, then it is weakly stationary since it has a finite variance. Conversely suppose X is Gaussian and weakly stationary. Then c(s, t) = cov(X(s), X(t)) depends
360
Solutions
Problems
[9.6.4]-[9.7.2]
on t - s only. The joint distribution of X(t1), X(t2), ... , X(tn) depends only on the common mean and the covariances c(ti, fj ). Now c(ti, fj) depends on fj - ti only, whence X (tl), X (t2), ... , X Un) have the same joint distribution as X(s + t1), X(s + t2), ... , X(s + tn). Therefore X is strongly stationary.
4.
(a) If s, t > 0, we have from Problem (4.14.13) that
whence cov(X(s) 2 , X(s
+ t) 2 ) = =
=
+ t) 2) - 1 m:{ X(s) 2 E(X(s + t) 2 I X(s))}- 1 c(t) 2E(X(s) 4 ) + (1- c(t) 2 )E(X(s) 2 )- 1 =
E(X(s) 2 X(s
by an elementary calculation. (b) Likewise cov(X(s) 3 , X(s + t) 3) = 3(3
2c(t) 2
+ 2c(t) 2 )c(t).
9.7 Solutions to problems
+ f3Yn-1, whence the autocovariance function c
1. It is easily seen that Yn = Xn +(a - f3)Xn-1 of Y is given by
if k = 0,
_1_+_a_2__ -=2-'-f3_2
1-/3
{ c(k) =
f31kl-1
{
2 }
a( 1 + af3- f3 ) 1- f32
if k
#
0.
Set Yn+ 1 = 2:::~0 ai Yn-i and find the ai for which it is the case that E{ (Yn+ 1 - Yn+ 1) Yn-k} = 0 for k ::::: 0. These equations yield 00
+ 1) =
c(k
L aic(k- i),
k::::: 0,
i=O
which have solution ai 2.
= a(f3- a)i fori
::::: 0.
The autocorrelation functions of X and Y satisfy r
aipx(n) = ap
L
ajakpy(n
+ k- j).
j,k=O
Therefore 2
aifx(A.)=
~;
oo
L
r
e-inJ...
n=-oo 2
L
ajakpy(n+k-j)
j,k=O oo
r
= ay ""'a·akei(k-j)J...
2n L
j,k=O
J
""' e-i(n+k-j)J...py(n+k-j)
L
n=-oo
2 iJ... 2 = ayiGa(e )I fy(A.).
361
[9.7.3]-[9.7.5]
Solutions
Stationary processes
In the case of exponential smoothing, G 0 (eil.) = (1 - t-t)/(1 - t-teil.), so that fx(A) =
c(l - t-t)2 fy(A) , 1- 2~-tcosA + t-t 2
IAI
< n,
1 is a constant chosen to make this a density function.
where c = a{; I a
Consider the sequence {Xn} defined by
3.
Xn
= Yn
- Yn
= Yn- aYn-1
- f3Yn-2·
Now Xn is orthogonal to {Yn-k : k 2: 1}, so that the Xn are uncorrelated random variables with spectral density function fx(A) = (2n)- 1 , A E (-n, n). By the result of Problem (9.7.2),
al fx (A) = a{; 11 -
aeil. - f3e 2il. 12 fy (A),
whence
f y (A)-
a2ja2 X
y
2n 11 -ae il. - f3 e 2il. 12 ,
-n
4. Let {X~ : n 2: 1} be the interarrival times of such a process counted from a time at which a meteorite falls. Then Xi, X~, ... are independent and distributed as X2. Let Y~ be the indicator function of the event {X:n = n for some m}. Then E(YmYm+n)
= lP'(Ym = 1, Ym+n = 1) = lP'(Ym+n = 1 I Ym = l)lP'(Ym =
1)
= lP'(Y~ =
where a = lP'(Ym = 1). The autocovariance function of Y is therefore c(n) n 2: 0, and Y is stationary. The spectral density function of Y satisfies
l)a
=
a{lP'(Y~
=
· , c(n) = Re { 1 ~ · , 1 } . fy(A) = - 1 ~ L..J e-znA L..J eznAc(n)-2n n=-oo a(l- a) na(l- a) n=O 2n Now
00
00
n=O
n=O
L einl. y~ = L eiAT~ where T~
= Xi + X~ + · · · + X~; just check the non-zero terms.
when eil.
i=
1, where (jJ is the characteristic function of x2· It follows that jy(A)
5.
Therefore
=
1 n(l-a)
Re {
1 1-(jJ(A)
a . } - -1, - -1-e11. 2n
We have that E(cos(nU)) =
l
:n:
-
1
-:n: 2n
cos(nu) du = 0,
362
IAI
< n.
1) -a},
Solutions
Problems
[9.7.6]-[9.7.7]
for n :::: 1. Also E(cos(mU) cos(nU)) =
m:{i (cos[(m + n)U] + cos[(m -
n)U1)} = 0
if m =f. n. Hence X is stationary with autocorrelation function p(k) = 8ko, and spectral density function f(J....) = (2n)- 1 for IJ....I < n. Finally E{ cos(mU) cos(nU) cos(rU)}
=
im:{ (cos[(m + n)U] + cos[(m- n)U]) cos(rU)}
= i{p(m
+ n- r) + p(m- n- r)}
which takes different values in the two cases (m, n, r) = (1, 2, 3), (2, 3, 4).
6. (a) The increments of N during any collection of intervals {(u;, v;) : 1 :::0 i :::: n} have the same fdds if all the intervals are shifted by the same constant. Therefore X is strongly stationary. Certainly E(X(t)) = J....a for all t, and the autocovariance function is c(t) = cov(X(O), X(t)) = { 0 J....(a- t)
if t >a, ifO :::0 t :::0 a.
Therefore the autocorrelation function is
p(t)
=
{
0 1- itfai
if it I >a, if it I Sa,
which we recognize as thecharacteristicfunctionofthe spectral density f(J....) = {1-cos(aJ....)}j(an J.. . 2 ); see Problems (5.12.27b, 28a). (b) We have that E(X(t)) = 0; furthermore, for s :::0 t, the correlation of X(s) and X(t) is
1
1
2(J cov(X(s), X(t)) = 2(J cov(W(s)- W(s -1), W(t)- W(t -1))
= s- min{s, t - 1}- (s- 1) + 1 if s:::: t - 1, { s-t+1
(s- 1)
ift-1
This depends on t - s only, and therefore X is stationary; X is Gaussian and therefore strongly stationary also. The autocorrelation function is
p(h) =
{
0 1- lhl
if lhl :::: 1, if lhl < 1,
which we recognize as the characteristic function of the density function f(J....)
= (1
-cos J....)j(nJ.... 2 ).
7. We have from Problem (8.7.1) that the general moving-average process of part (b) is stationary with autocovariance function c(k) = '£}=o a}ak+ J, k :::: 0, with the convention that as = 0 if s < 0 ors > r. (a) In this case, the autocorrelation function is
~
p(k)={ 1 +a 2 0
363
if k = 0, if lkl = 1, iflk1>1,
[9.7.8]-[9.7.13]
Solutions
Stationary processes
whence the spectral density function is 1 ( p(O) + e 1·;.. p(l) + e- 1·;.. p(-1) ) = 1 ( 1+ 2acosA.) f(A.) = 2 , ~ ~ 1+a
IA.I < n.
(b) We have that
/(A.)=__!__
00
L
e-ik)..p(k)
=
2n k=-oo
8.
=
')..
IA(ez )I 2nc(O)
2
"2:,1 a1zi. See Problem (9.7.2) also. The spectral density function f is given by the inversion theorem (5.9.1) as
where c(O) =
"2:,1 aJ and A(z)
_1_'La·eij).. Lak ·e-i(k+j)).. 2nc(O) J 1 +J k
=
f(x)
= -1
100 e-ztx . p(t) dt
2n _ 00
00 lp(t)l dt < oo; see Problem (5.12.20). Now 1 100 lp(t)l dt 1/(x)l .:5-
under the condition J0
2n -oo
and
lf(x +h)- f(x)l .:5
__!__
100
2n -oo
Ieith- 11· lp(t)l dt.
The integrand is dominated by the integrable function 21p(t)l. Using the dominated convergence theorem, we deduce that lf(x +h)- f(x)l--+ 0 ash--+ 0, uniformly in x. By Exercise (9.5.2), var(n- 1 'L,'J=l XJ) --+ a 2 if Cn = n- 1 'L-j:J cov(Xo, Xj) --+ a 2 . If cov(Xo, Xn) --+ 0 then Cn --+ 0, and the result follows.
9.
10. Let X 1, X 2, ... be independent identically distributed random variables with mean p,. The sequence X is stationary, and it is a consequence of the ergodic theorem that n- 1 'L,'J=I 1 --+ Z a.s. and in mean, where Z is a tail function of X 1, X 2• ... with mean p,. Using the zero-one law, Z is a.s. constant, and therefore lP'(Z = p,) = 1.
x
11. We have from the ergodic theorem that n- 1 'L-t=l Yi --+ lE(Y I 1) a.s. and in mean, where 1 is the a-field of invariant events. The condition of the question is therefore lE(Y 11) = lE(Y)
a.s., for all appropriate Y.
Suppose(*) holds. Pick A E 1, and set Y = lA to obtain lA = Q(A) a.s. Now lA takes the values 0 and 1, so that Q(A) equals 0 or 1, implying that Q is ergodic. Conversely, suppose Q is ergodic. Then lE(Y I 1) is measurable on a trivial a-field, and therefore equals lE(Y) a.s.
12. Suppose Q is strongly mixing. If A is an invariant event then A = r-nA. Therefore Q(A) = Q(A n r-nA) --+ Q(A) 2 as n --+ oo, implying that Q(A) equals 0 or 1, and therefore Q is ergodic.
13. The vector X = (X 1, X2, ... ) induces a probability measure Q on (l~T, JF,T). Since Tis measurepreserving, Q is stationary. Let Y : rntT --+ rnt be given by Y(x) = x1 for x = (xl, x2, .. .), and define Yi (x) = Y(ri-l (x)) where r is the usual shift operator on rntT. The vector Y = (f1, Y2, ... ) has the same distributions as the vector X. By the ergodic theorem for Y, n- 1 'L-t=l Yi --+ lE(Y I 'J.) a.s. and in mean, where '!f. is the invariant a -field of r. It follows that the limit 1 n Z = lim - ""Xi n--+oo n L...t i=l
364
Solutions
Problems
exists a.s. and in mean. Now U = limsupn-+oo (n- 1
[9. 7.14]-[9. 7.15]
L;1 Xi) is invariant, since
implying that U(w) = U(Tw) a.s. It follows that U is 1-measurable, and it is the case that Z = U a.s. Take conditional expectations of(*), given 1, to obtain U = E(X 11) a.s. If Tis ergodic, then 1 is trivial, so that E(X 11) is a.s. constant; therefore E(X 11) = E(X) a.s.
14. If(a,b) ~ [0, l),thenT- 1(a,b) = (1a, ib)u(i+1a, 1+!-b),andthereforeTismeasurable. Secondly, 1P'(T- 1(a, b))= 2(1b -1a) = b- a= IP'((a, b)), so that r- 1 preserves the measure of intervals. The intervals generate :B, and it is then standard that r- 1 preserves the measures of all events. Let A be invariant, in that A= r- 1A. Let 0 :S: w < 1; it is easily seen that T(w) = T(w + 1). Therefore wE A if and only if w + 1 E A, implying that An [1, 1) = lP'(A n E)= 11P'(A) = lP'(A)lP'(E)
i + {An [0, 1) }; hence
forE= [0, 1), [1, 1).
This proves that A is independent of both [0, 1) and [1, 1). A similar proof gives that A is independent of any set E which is, for some n, the union of intervals of the form [k2-n, (k+ 1)2-n) forO ::::: k < 2n. It is a fundamental result of measure theory that there exists a sequence E 1, Ez, ... of events such that (a) En is of the above form, for each n, (b) lP'(A LEn) --+ 0 as n --+ oo. Choosing the En accordingly, it follows that lP'(A n En) = lP'(A)lP'(En) --+ lP'(A) 2
by independence,
IIP'(A n En) -lP'(A)I ::::: lP'(A LEn) --+ 0. Therefore IF'( A) = lP'(A) 2 so that lP'(A) equals 0 or 1. For w E Q, expand w in base 2, w = O.w1 wz · · · , and define Y (w) = w1. It is easily seen that Y(Tn- 1w) = Wn, whence the ergodic theorem (Problem (9.7.13)) yields that n- 1 L;?=l Wi --+ as n --+ oo for all w in some event of probability 1.
i
15. We may as well assume that 0 < a < 1. LetT : [0, 1) --+ [0, 1) be given by T(x) = x +a (mod 1). It is easily seen that Tis invertible and measure-preserving. Furthermore T(X) is uniform on [0, 1], and it follows that the sequence Z 1, Zz, ... has the same fdds as Zz, Z3, ... , which is to say that Z is stationary. It therefore suffices to prove that T is an ergodic shift, since this will imply by the ergodic theorem that 1 n {I Zj --+ E(ZI) = g(u) du. n j=l o
-L
Jo
We use Fourier analysis. Let A be an invariant subset of [0, 1). The indicator function of A has a Fourier series: 00
IA(x) ~
L
anen(x)
n=-oo
where en (x) = e 2:rrinx and
1 an=-
2n
lolo
IA(x)e-n(x)dx = 1 -
2n
365
1
e-n(x)dx.
A
[9.7.16]-[9.7.18]
Solutions
Stationary processes
Similarly the indicator function of r- 1 A has a Fourier series, lr-!A(x) ~ Lbnen(x) n
where, using the substitution y
= T(x),
since em(Y- a)= e- 2nimaem(y). Therefore Ir-IA has Fourier series I r-IA (x ) ~ "" L..Je -2rrina anen ( x ) . n
Now I A = I r-1 A since A is invariant. We compare the previous formula with that of (* ), and deduce that an = e- 2 nina an for all n. Since a is irrational, it follows that an = 0 if n =f 0, and therefore I A has Fourier series ao, a constant. Therefore lA is a.s. constant, which is to say that either lP'(A) = 0 or lP'(A) = 1.
16. Let G 1 (z) = E(zX(t)), the probability generating function of X(t). Since X has stationary independent increments, for any n (~ 1), X(t) may be expressed as the sum n
X(t) = L { X(itjn)- X((i -l)tjn)} i=l
of independent identically distributed variables. Hence X(t) is infinitely divisible. By Problem (5.12.13), we may write Gt(z)
= e-J..(t)(l-A(z))
for some probability generating function A, and some A.(t). Similarly, X(s + t) = X(s) +{X (s + t)- X(s)}, whence Gs+t(z) = Gs(z)G 1 (z), implying that G 1 (z) = e!h(z)t for some t-t(z); we have used a little monotonicity here. Combining this with(*), we obtain that G 1 (z) = e-J..t(l-A(z)) for some A.. Finally, X (t) has jumps of unit magnitude only, whence the probability generating function A is given by A(z) = z.
17. (a) We have that X(t)- X(O)
= {X(s)- X(O)}
+ { X(t)- X(s) },
0:::::
s::::: t,
whence, by stationarity, {m(t) -m(O)} = {m(s)- m(O)} + {m(t -s) -m(O)}.
Now m is continuous, so that m(t)- m(O) = {Jt, t ~ 0, for some {3; see Problem (4.14.5). (b) Take variances of(*) to obtain v(t) = v(s) + v(t- s), 0 ::::: s ::::: t, whence v(t) = a 2 t for some (J2.
18. In the context of this chapter, a process Z is a standard Wiener process if it is Gaussian with Z(O) = 0, with zero means, and autocovariance function c(s, t) = min{s, t}.
366
Solutions [9.7.19]-[9.7.20]
Problems
(a) Z(t)
= aW(tja 2 ) satisfies Z(O) = 0, JE(Z(t)) = 0, and = a 2 min{s ja 2 , t ja 2 } = min{s, t}.
cov(Z(s), Z(t))
(b) The only calculation of any interest here is cov(W(s + a)-W(a), W(t +a)- W(a))
= c(s +a, t +a) -
c(a, t +a) - c(s +a, a)+ c(a, a)
=(s+a)-a-a+a=s,
s~t.
(c) V(O) = 0, and JE(V(t)) = 0. Finally, if s, t > 0, cov(V(s), V(t)) = stcov(W(l/s), W(lft)) = stmin{l/s, 1ft}= min{t, s}. (d) Z(t)
=
W(l)- W(l - t) satisfies Z(O) cov(Z(s), Z(t))
= 0, JE(Z(t)) = 0.
Also Z is Gaussian, and
= 1- (1- s)- (1- t) + min{l- s, 1- t} = min{s, t}, O~s,t~l.
19. The process W has stationary independent increments, and G(t) = JE(IW(t)l 2 ) satisfies G(t) t --+ 0 as t --+ 0; hence ¢ (u) d W (u) is well defined for any ¢ satisfying
jgo
=
fooo l¢(u)l 2 dG(u) = fooo ¢(u) 2 du < oo. It is obvious that ¢(u) = l[O,tJ(u) and ¢(u) = e-(t-u) l[O,tJ(u) are such functions. Now X(t) is the limit (in mean-square) of the sequence n-1
Sn(t)
= l:)W((j + 1)t/n)- W(jtjn)},
n~l.
}=0
However Sn(t) = W(t) for all n, and therefore Sn(t) ~ W(t) as n--+ oo. Finally, Y (s) is the limit (in mean-square) of a sequence of normal random variables with mean 0, and therefore is Gaussian with mean 0. If s < t, cov(Y(s), Y(t)) = fooo (e-(s-u) l[O,sJ(u)) (e-(t-u) I[o,t](u)) dG(u) -los e 2u-s-t d u_- 21 ( e s-t -e -s-t) . 0
Y is an Omstein-Uhlenbeck process. 20. (a) W(t) is N(O, t), so that JEIW(t)l =
l
oo
-00
var(IW(t)l)
lui
1
2
--e-z
..;z;rt
2)
= JE(W(t) 2 ) - 2t n = t ( 1-; 367
J2ili(, .
[9. 7.21]-[9. 7.21]
Solutions
Stationary processes
The process X is never negative, and therefore it is not Gaussian. It is Markov since, if s < t and B is an event defined in terms of {X(u) : u ::5 s}, then the conditional distribution function of X(t) satisfies
I
lP'(X(t) ::5 y X(s)
= x, B) = lP'(X(t) ::5 y IW(s) = x, B)lP'(W(s) =xI X(s) = x, B) + lP'(X(t) ::5 y IW(s) = -x, B)lP'(W(s) =-xI X(s) = x, B) = i{lP'(X(t) ::5 y IW(s) =x) +lP'(X(t) ::5 y IW(s) = -x) }.
which does not depend on B. (b) Certainly, E(Y(t))
=
1 Jiiit 00
I
eU
2
--e-z(u ft) du
I
= e2 1 •
-00
Secondly, W(s) + W(t) = 2W(s) + {W(t)- W(s)} is N(O, 3s + t) if s < t, implying that E(Y(s)Y(t)) = JE: (eW(s)+W(t)) = A(3s+t), and therefore
I
I
cov(Y(s), Y(t)) = e'2( 3s+t)- ez(s+t),
s < t.
W(1) is N(O, 1), and therefore Y(1) has the log-normal distribution. Therefore Y is not Gaussian. It is Markov since W is Markov, and Y(t) is a one-{Jne function of W (t). (c) We shall assume that the random function W is a.s. continuous, a point to which we return in Chapter 13. Certainly, E(Z(t)) =lot E(W(u)) du E(Z(s)Z(t))
=
!1
= 0,
E(W(u)W(v))dudv
O
o;v;t
=1s u=O
{fu vdv+ t udv}du=is (3t-s), lv=O lv=u 2
s
since E(W(u)W(v)) = min{u, v}. Z is Gaussian, as the following argument indicates. The single random variable Z(t) may be expressed as a limit of the form lim
n-+oo
~ (!_) W(itjn), n
L...J
i=l
each such summation being normal. The limit of normal variables is normal (see Problem (7.11.19)), and therefore Z(t) is normal. The limit in(*) exists a.s., and hence in probability. By an appeal to (7.11.19b), pairs (Z(s), Z(t)) are bivariate normal, and a similar argument is valid for all n-tuples of the Z(u). The process Z is not Markov. An increment Z (t) - Z (s) depends very much on W (s) = Z' (s ), and the collection {Z (u) : u ::; s} contains much information about Z' (s) in excess of the information contained in the single value Z(s).
=
=
=
=
21. Let Ui X(ti). The random variables A= U1, B Uz- U1o C U3- Uz, D U4- U3 are independent and normal with zero means and respective variances t1, tz - t1, t3 - tz, t4 - t3. The Jacobian of the transformation is 1, and it follows that U1, Uz, U3, U4 have joint density function
e-!Q /u(u) = (2n) 2,Jtl(t2- t1)(t3- tz)(t4- t3)
368
Solutions
Problems
[9. 7.22]-[9. 7.22]
where ui
(u2- u1) 2
fl
f2 - tl
Q=-+ Likewise ul and
+
(u3 - u2) 2 t3 - t2
(u4- u3) 2
+---'--~
f4 - t3
u4 have joint density function where
Hence the joint density function of U2 and U3, given U1 = U4 = 0, is
where
)2 2 2 ( S=~+ U3-U2 +~. f2 - fl t3 - t2 f4 - f3
Now g is the density function of a bivariate normal distribution with zero means, marginal variances (t2 - tl )(t3 - t2) t3- tl
and correlation
See also Exercise (8.5.2). 22. (a) The random variables {/j (x) : I :::: j :::: n} are independent, so that E(Fn(x)) =X,
1 x(l-x) var(Fn(x)) = -var(h(x)) = . n n
By the central limit theorem, ,Jn{Fn(x)- x} ~ Y(x), where Y(x) is N(O, x(1- x)). (b) The limit distribution is multivariate normal. There are general methods for showing this, and here is a sketch. If 0 ::::= x1 < x2 ::::= 1, then the number M2 (= nF(x2)) of the Ij not greater than x2 is approximately N(nx2, nx2(l- x2)). Conditional on {M2 = m}, the number M1 = nF(xl) is approximately N(mu, mu(l - u)) where u = xlfx2. It is now a small exercise to see that the pair (M1, M2) is approximately bivariate normal with means nx1, nx2, with variances nx1 (1 - x1), nx2(l - x2), and such that
whence cov(Ml, M2) ~ nx1 (1 - x2). It follows similarly that the limit of the general collection is multivariate normal with mean 0, variances x;(1- x;), and covariances Cij = x;(l- Xj). (c) The autocovariance function of the limit distribution is c(s, t) = min{s, t} - st, whereas, for 0:::: s :::: t :::: 1, we have that cov(Z(s), Z(t)) = s- ts- st + st = min{s, t}- st. It may be shown that the limit of the process { Jn (Fn (x) - x) : n :::: 1} exists as n -+ oo, in a certain sense, the limit being a Brownian bridge; such a limit theorem for processes is called a 'functional limit theorem'.
369
10 Renewals
10.1 Solutions. The renewal equation 1.
SincelE(Xl) > 0, there exists E (> 0) such thatlP'(X1 2: E)> E. LetXIc = El{xk::::E}• and denote
by N' the related renewal process. Now N(t) :s N'(t), so that lE(eeN(t)) :s lE(eeN'{tl), fore > 0. Let Zm be the number of renewals (in N') between the times at which N' reaches the values (m - 1)E and mE. The Z's are independent with Eee JE(ee Zm) _ ----~ - 1 - (1 - E)ee'
if (I- E)ee
whence JE(eeN'(t)) :S (Eee {1- (1- E)ee} - 1)t/E for sufficiently small positive e.
2. Let X 1 be the time of the first arrival. If X 1 > s, then W = s. On the other hand if X 1 < s, then the process starts off afresh at the new starting time X 1. Therefore, by conditioning on the value of
x1, Fw(x) =
fooo lP'(W :s
=los
X
I x1
=los
= u)dF(u)
lP'(W :S x- u)dF(u)
lP'(W
:s X - u)dF(u) +
1
00
1· dF(u)
+ {1- F(s)}
if x 2: s. It is clear that F w (x) = 0 if x < s. This integral equation for F w may be written in the standard form Fw(x) = H(x)
where H and
+fox Fw(x- u)dF(u)
F are given by H(x) = {
~- F(s)
if X <
S,
ifx:=::s,
~
F(x) =
{ F(x)
F(s)
if x < s, if X 2: s.
This renewal-type equation may be solved in the usual way by the method ofLaplace-Stieltjes transforms. We have that Fw = H* + FiVF*, whence Fw = H* j(l- F*). If N is a Poisson process then F(x) = 1- e-l.x. In this case H*(e) =
fooo e-ex dH(x) = e-(.A.+e)s,
since H is constant apart from a jump at x = s. Similarly
370
Solutions
Limit theorems
so that
*
(A.+ e)e-(A+O)s
Fw(e)
Finally, replace e with 3.
[10.1.3]-[10.2.2]
=
e
+ A.e-(A+O)s
.
-e, and differentiate to find the mean.
We have as usual that lP'(N(t) = n) = lP'(Sn :S t) -lP'(Sn+1 :S t). In the respective cases,
(a)
- - Jor {;,.nbxnb-1
(b)
4.
lP'(N(t)- n)-
r(nb)
-
;,.(n+1)bx(n+1)b-1} -A.x f((n + l)b) e dx.
By conditioning on X1, m(t) = E(N(t)) satisfies m(t) =lot (I+ m(t- x)) dx = t +lot m(x) dx,
O:::;t:::;l.
Hence m' = 1 + m, with solution m(t) = e1 - 1, for 0 ::::; t ::::; 1. (For larger values oft, m(t) = 1 + m(t- x) dx, and a tiresome iteration is in principle possible.)
JJ
With v(t) = E(N(t) 2 ), v(t)= l[v(t-x)+2m(t-x)+l]dx=t+2(e 1 -t-l)+ lot v(x)dx,
Hence v' = v + 2et- 1, with solution v(t) = 1 - et
+ 2te1 for 0::::; t ::::;
O:::;t:::;l.
1.
10.2 Solutions. Limit theorems 1. Let Zi be the number of passengers in the ith plane, and assume that the Zi are independent of each other and of the arrival process. The number of passengers who have arrived by time t is S(t) = I:~i) Z;. Now 1 N(t) S(t) E(Z1) -S(t) = - - · ----+ - a.s. t t N(t) f.l
by the law of the large numbers, since N(t)/t --+ 2.
1/Jl a.s., and N(t)
--+ oo a.s.
We have that
since l(M:;:i)l{M:;:JJ = /{M:;:iv j}• where i V j = max{i, j}. Now
since {M :S i - 1} is defined in terms of Z 1, Z2, ... , Zi -1, and is therefore independent of Z;. Similarly E(Z; ZJ l(M:;:jj) = E(ZJ )E(Zi l(M:;:jj) = 0 if i < j. It follows that E(Tli) =
00
00
i=1
i=1
2:: E(Zf)lP'(M 2:: i) = a 2 2:: lP'(M 2:: i) = a 2E(M). 371
[10.2.3]-[10.2.5]
Solutions
Renewals
3. (i) The shortest way is to observe that N(t) + k is a stopping time if k :::: 1. Alternatively, we have by Wald's equation that JE(TN(t)+I) = Jl(m(t) + 1). Also
I
lE(XN(t)+k) = lE{lE(XN(t)+k N(t))} =
k:::: 2,
jl,
and therefore, for k :::: 1, k
lE(TN(t)+k) = lE(TN(t)+I)
+ 2:JE(XN(t)+j) =
Jl(m(t)
+ k).
j=2
(ii) Suppose p
=f: 1 and li"(Xt=a)= {
Then f.l = 2 - p
=f: 1.
p
if a= 1,
1- p
if a= 2.
Also
JE(TN(I)) = (1- p)JE(To 1 N(1) = O) whereas m(1) = p. Therefore JE(TN(I))
=f:
+ plE(Tt
1
N(1) = 1) = p,
Jlm(l).
4. Let V(t) = N(t) + 1, and let W1, W2, ... be defined inductively as follows. W1 = V(1), W2 is obtained similarly to Wt but relative to the renewal process starting at the V(1)th renewal, i.e., at time TN(I)+1· and Wn is obtained similarly: Wn = N(Txn-l
+ 1)- N(Txn_ 1 ) + 1,
n:::: 2,
whereXm = W1 + W2+· · ·+ Wm. Foreachn, Wn isindependentofthesequence W1, W2, ... , Wn-1• and therefore the Wn are independent copies of V ( 1). It is easily seen, by measuring the time-intervals covered, that V(t).:::; 2::[~ 1 W;, and hence 1 1 ftl - V(t).:::; -2: Wi--+ lE(V(l))
t
a.s. and in mean, as t--+ oo.
t i=I
It follows that the family { m- 1 2::~ 1 W; : m :::: 1} is uniformly integrable (see Theorem (7.10.3)). Now N(t).:::; V(t), and so {N(t)/t: t:::: 0} is uniformly integrable also. Since N (t) It ~ J.l- 1, it follows by uniform integrability that there is also convergence in mean.
5.
(a) Using the fact that lP'(N(t) = k) = lP'(Sk .:::; t) -li"(Sk+1 .:::; t), we find that lE(sN(T)) =
=
fooo (~ sklP'(N(t) = k)) ve-vt dt ~ sk {fooo [li"(Sk .:::; t) -lP'(Sk+I .:::; t)] ve-vt dt} .
By integration by parts, J000 lP'(Sk.:::; t)ve-vt dt = M(-v)k fork:::: 0. Therefore, lE(sN(T)) = f>k{M(-v)k _ M(-v)k+I} = 1- M(-v). k=O 1- sM(-v) (b) In this case, lE(sN(T)) = lE(eAT(s-l)) = MT(A.(s -1)). When T has the given gamma distribution, MT(e) = {vj(v- e)}b, and
lE(sN(T)) =
(-v-) (1 - ~) b
v+A.
v+A.
The coefficient of sk may be found by use of the binomial theorem.
372
b
Excess life
Solutions
[10.3.1]-[10.3.3]
10.3 Solutions. Excess life 1. Let g (y) = lP'( E (t) > y), assumed notto depend on t. By the integral equation for the distribution of E(t),
l
+ y) + g(y) dF(x). Write h(x) = 1- F(x) to obtain g(y)h(t) = h(t + y), for y, t :=:: 0. With t = 0, we have that g(y)h(O) = h(y), whence g(y) = h(y)/ h(O) satisfies g(t + y) = g(t)g(y), for y, t :=:: 0. Now g is g(y)
= 1-
F(t
left-continuous, and we deduce as usual that g(t) the renewal process is a Poisson process.
= e-At for some A..
Hence F(t)
=
1 -e-At, and
2.
(a) Examine a sample path of E. If E(t) = x, then the sample path decreases (with slope -1) until it reaches the value 0, at which point it jumps to a height X, where X is the next interarrival time. Since X is independent of all previous interarrival times, the process is Markovian. (b) In contrast, C has sample paths which increase (with slope 1) until a renewal occurs, at which they drop to 0. If C (s) = x and, in addition, we know the entire history of the process up to time s, the time of the next renewal depends only on the length ofthe spent period (i.e., x) of the interarrival time in process. Hence C is Markovian.
3.
(a) We have that lP'(E(t) _:::: y) = F(t
+ y)
-l
G(t
+ y- x)dm(x)
where G(u) = 1- F(u). Check the conditions of the key renewal theorem (10.2.7): g(t) = G(t + y) satisfies: (i) g(t) :::: 0, (ii) f0 g(t) dt .:::= fcF[1 - F(u)] du = E(X 1) < oo, (iii) g is non-increasing. We conclude, by that theorem, that
00
1 lim lP'(E(t) _:::: y) = 1 - -
t-+00
f1
100 g(x)dx = 0
loy -[11 F(x)]dx. 0 f1
(b) Integrating by parts,
i oo o
xr 1 -[1-F(x)]dx=11
11
laoo
xr+1 E(xr+1) --dF(x)= 1 . M(r + 1)
o r+1
See Exercise (4.3.3). (c) As in Exercise (4.3.3), we have that E(E(tY) = J0 ryr- 11P'(E(t) > y) dy, implying by(*) that
00
E(E(tn = E({CX1- t)+n
+ i:ol~ 0 ryr- 1 1P'(X 1
> t
+ y -x)dm(x)dy,
whence the given integral equation is valid with
Now h satisfies the conditions of the key renewal theorem, whence lim E(E(tY)
t-+oo
= .!._ { 00 h(u) du = .!._ { { yr dF(u + y) du 11 Jo 11 JJoy)dy= . 11 o J1(r + 1) 373
[10.3.4]-[10.3.5]
4.
Renewals
Solutions
We have that lP'(E(t)>yjC(t)=x)=lP'(Xt>y+xiXt>x)=
1- F(y +x)
1- F(x)
,
whence E(E(t)jC(t)=x)= fool-F(y+x)dy=E{(Xt-x)+}_ Jo 1- F(x) 1- F(x)
5. (a) Apply Exercise (10.2.2) to the sequence X;- f-L, 1 :::= i < oo, to obtain var(TM(t)- t-LM(t)) = a 2 E(M(t)). (b) Clearly TM(t) = t + E(t), where E is excess lifetime, and hence t-LM(t) = (t + E(t))- (TM(t)t-LM(t)), implying in tum that f-L 2 var(M(t)) =
where
SM(t)
var(E(t)) + var(SM(t))- 2cov(E(t),
SM(t)),
= TM(t)- {LM(t). Now
as t --+ oo
ifE(XI) < oo (see Exercise (10.3.3c)), implying that 1 - var(E(t)) --+ 0 t
as t --+ oo.
This is valid under the weaker assumption that E(XI) < oo, as the following argument shows. By Exercise (10.3.3c), E(E(t) 2 ) = a(t) +lot a(t- u) dm(u), where a(u) = JE:({(Xt- u)+} 2 ). Now use the key renewal theorem together with the fact that a(t):::: E(XIl(x 1 >tJ)--+ Oast--+ oo. Using the Cauchy-Schwarz inequality,
as t --+ oo, by part (a) and (** ). Returning to (*), we have that f-L2 -var(M(t))--+ lim t t--'>00
{a2
-(m(t)+l) } =a2 -. t {L
374
Solutions [10.4.1]-[10.5.3]
Renewal-reward processes
10.4 Solution. Applications 1. Visualize a renewal as arriving after two stages, type 1 stages being exponential parameter A and type 2 stages being exponential parameter J-t. The 'stage' process is the flip-flop two-state Markov process of Exercise (6.9.1). With an obvious notation, Pn(t)
= _A_e-(A+JL)t + _1-t_. A+J-t
A+J-t
Hence the excess lifetime distribution is a mixture of the exponential distribution with parameter J-t, and the distribution of the sum of two exponential random variables, thus,
where g(x) is the density function of a typical interarrival time. By Wald's equation, E(t
+ E(t)) = E(SN(t)+1) = E(X1)E(N(t) + 1) =
We substitute
(± + ~)
(m(t)
+ 1).
p 11 (t) (-A1+-1-t1) + (1- Pn(t))-1-t1=1-t-1 + A
E(E(t)) = Pn(t) to obtain the required expression.
10.5 Solutions. Renewal-reward processes 1. Suppose, at time s, you are paid a reward at rate u(X(s)). By Theorem (10.5.10), equation (10.5.7), and Exercise (6.9.11b),
lot
1 t 0
-
Suppose lu (i) I .::; K < oo for all i
E
a.s.
1 J-tjgj
l{X(s)=j)dS ~ - - = lfj.
S, and let F be a finite subset of the state space. Then
where T1 (F) is the total time spent in F up to time t. Take the limit as t F t S, to obtain the required result.
~
oo using (*), and then as
2. Suppose you are paid a reward at unit rate during every interarrival time of type X, i.e., at all times tat which M(t) is even. By the renewal-reward theorem (10.5.1), 1
t
lot I
a.s. E(reward during interarrival time) EX 1 = . E(length of interarrival time) EX 1 + EY1
M s is even ds ~
o { ()
}
3.
Suppose, at timet, you are paid a reward at rate C(t). The expected reward during an interval (cycle) of length X is fox s ds = ~ X 2 , since the age C is the same at the time s into the interval. The
375
[10.5.4]-[10.6.3]
Solutions
Renewals
result follows by the renewal-reward theorem (10.5.1) and equation (10.5.7). The same conclusion is valid for the excess lifetime E(s), the integral in this case being fox (X- s) ds = ~ X 2 . 4. Suppose Xo = j. Let V1 = min{n ::: 1 : Xn = j, Xm = k for some 1 :S m < n}, the first visit to j subsequent to a visit to k, and let Vr+1 = min{n ::: Vr : Xn = j, Xm = k for some Vr + 1 :S m < n}. The Vr are the times of a renewal process. Suppose a reward of one ecu is paid at every visit to k. By the renewal-reward theorem and equation (10.5.7),
By considering the time of the first visit to k, E(V1
I Xo = j) = E(Tk I Xo = j) +E(1) I Xo = k).
The latter expectation in (*) is the mean of a random variable N having the geometric distribution f'(N = n) = p(l - p)n- 1 for n ::: 1, where p = f'(1) < Tk I Xo = k). Since E(N) = p- 1, we deduce as required that 1/f'(1) < Tk I Xo = k) lfk = - - - - " - - - - - - - - -
E(n
I Xo
= j) + E(1j
I Xo
= k)
10.6 Solutions to problems 1. (a) For any n, JFl'(N(t) < n) :S JFl'(Tn > t)--+ 0 as t--+ oo. (b) Either use Exercise (10.1.1), or argue as follows. Since JL > 0, there exists E (> 0) such that f'(X1 >E) > 0. For all n, f'(Tn :S nE) = 1- f'(Tn > nE) :S 1- f'(X1 > E)n < 1,
so that, if t > 0, there exists n = n(t) such that f'(Tn :S t) < 1. Fix t and let n be chosen accordingly. Any positive integer k may be expressed in the form k =an+ fj where 0 :S fj < n. Now JFl'(Tk :S t) :S f'(Tn :S t)a for an :S k < (a+ 1)n, and hence 00
00
k=1
a=O
L JFl'(Tk :S t) :S L nf'(Tn :S t)a <
m(t) =
00.
(c) It is easiest to use Exercise (10.1.1), which implies the stronger conclusion that the moment generating function of N (t) is finite in a neighbourhood of the origin. 2.
(i) Condition on X 1 to obtain
r
,, 2 v(t)= Jo E{(N(t-u)+1) }dF(u)= Jo {v(t-u)+2m(t-u)+1}dF(u).
Take Laplace-Stieltjes transforms to find that v* = (v* +2m*+ 1)F*, where m* = F* + m* F* as usual. Therefore v* = m*(l +2m*), which may be inverted to obtain the required integral equation. (ii) If N is a Poisson process with intensity A., then m(t) =At, and therefore v(t) 3.
Fix x
E
lit Then
376
= (A.t) 2 + A.t.
Solutions [10.6.4]-[10.6. 7]
Problems wherea(t) = l
lP'(T () < t) = lP' ( Ta(t) - p,a(t) < t - p,a(t)) a t a ..j(i(i) - a ..j(i(i) . However a(t)
~
tIp, as t --+ oo, and therefore t - p,a(t)
---'==='- --+ a ..j(i(i)
as t--+ oo,
-X
implying by the usual central limit theorem that lP'(N(t)-(t/p,) y'ta2jp,3
~x)
__... <1>(-x)
as t--+ oo
where is the N(O, 1) distribution function. An alternative proof makes use of Anscombe's theorem (7.11.28). 4.
We have that, for y .::0 t, lP'(C(t) ~ y) = lP'(E(t- y) > y)--+ lim lP'(E(u) > y) U---+00
as t--+ oo
= {oo !._[1- F(x)] dx
}y
1-i
by Exercise (10.3.3a). The current and excess lifetimes have the same asymptotic distributions.
5. Using the lack-of-memory property of the Poisson process, the current lifetime C(t) is independent of the excess lifetime E(t), the latter being exponentially distributed with parameter J..... To derive the density function of C(t) either solve (without difficulty in this case) the relevant integral equation, or argue as follows. Looking backwards in time from t, the arrival process looks like a Poisson process up to distance t (at the origin) where it stops. Therefore C(t) may be expressed as min{Z, t} where Z is exponential with parameter J....; hence J....e-J..s
!qt)(s) = { 0
and lP'(C(t) = t) = e-At. Now D(t) convolution formula) to be as given.
=
C(t)
if s < t ·f - , s > t,
1
+ E(t),
whose distribution is easily found (by the
6. The ith interarrival time may be expressed in the form T + Zi where Zi is exponential with parameter J..... In addition, Z1, Zz, ... are independent, by the lack-of-memory property. Now
1- F(x)
= lP'(T + zl
> x)
= lP'(Zl
>X- T)
= e-A(x-T)'
X~
T.
Taking into account the (conventional) dead period beginning at time 0, we have that
lP'(N(t)
~ k)
k
= lP'(kT
+ ~ Zi
.::0 t) = lP'(N(t- kT)
~ k),
t
~
kT,
l=l
where N is a Poisson process. 7. We have that X1 = L + E(L) where Lis the length of the dead period beginning at 0, and E(L) is the excess lifetime at L. Therefore, conditioning on L,
377
[10.6.8]-[10.6.10]
Solutions
Renewals
We have that lP'(E(t) s y) = F(t
+ y)
-lot {1-
F(t
+ y -x)}dm(x).
By the renewal equation,
m(t
+ y) =
F(t
rr+y F(t + y- x)dm(x),
+ y) + Jo
whence, by subtraction,
i
t+y
{1-F(t+y-x)}dm(x).
lP'(E(t)sy)= t It follows that
1P'(X 1 sx)=
{x
{x {1-F(x-y)}dm(y)dFL(l)
}z=O }y=l
= [FL(l) {x{1-F(x-y)}dm(y)]x
~0
h
+
r FL(l){1-F(x-l)}dm(l)
k
using integration by parts. The term in square brackets equals 0.
8. (a) Each interarrival time has the same distribution as the sum of two independent random variables with the exponential distribution. Therefore N(t) has the same distribution as L~ M(t)J where M is a Poisson process with intensity A. Therefore m(t) = ~JE(M(t))- ilP'(M(t) is odd). Now JE(M(t)) = At, and oo (At)Zn+le-At lP'(M(t) is odd)="'"' = ~e-A.t(eA.t- e-At). ~ (2n + 1)!
n=O
With more work, one may establish the probability generating function of N(t). (b) Doing part (a) as above, one may see that iii(t) = m(t).
9.
Clearly C(t) and E(t) are independent if the process N is a Poisson process. Conversely, suppose that C(t) and E(t) are independent, for each fixed choice oft. The event {C(t) 0:: y} n {E(t) 0:: x} occurs if and only if E (t - y) ::=: x + y. Therefore lP'(C(t) ::=: y)lP'(E(t) ::=: x) = lP'(E(t- y) ::=: x
+ y).
Take the limit as t ~ oo, remembering Exercise (10.3.3) and Problem (10.6.4), to obtain that G(y)G(x) = G(x + y) if x, y ::=: 0, where
1
00
G(u) =
u
1
-[1- F(v)] dv. f.-i
Now 1- G is a distribution function, and hence has the lack-of-memory property (Problem (4.14.5)), implying that G(u) = e-Jcu for some A. This implies in tum that [1- F(u)]/ f.-i = -G'(u) = Ae-Jcu, whence f.-i = 1/A and F(u) = 1- e-Jcu.
10. Clearly N is a renewal process if Nz is Poisson. Suppose that N is a renewal process, and write A for the intensity of N 1 , and Fz for the interarrival time distribution of Nz. By considering the time X 1 to the first arrival of N,
378
Solutions [10.6.11]-(10.6.12]
Problems
Writing E, Ei for the excess lifetimes of N, Ni, we have that
Take the limit as t
~
oo, using Exercise (10.3.3), to find that
1 100 -[1-
F(u)] du = e-Ax
xf.L
100 -[11 F2(u)]
du,
xf.L2
where f.L2 is the mean of F2. Differentiate, and use(*), to obtain
which simplifies to give 1- F2(x) = c equation has solution F2 (x) = 1 - e -ex.
fx
00
[1- F2(u)] du where c = AJL/(JL2- J.L); this integral
11. (i) Taking transforms of the renewal equation in the usual way, we find that m* e -
F*(e)
< ) - 1- F*(e)
where as
-----1 1- F*(e)
F*(e) = E(e-ex,) = 1- elL+ ~e 2 (JL 2
+a 2 ) +o(e 2 )
e ~ 0. Substitute this into the above expression to obtain
and expand to obtain the given expression. A formal inversion yields the expression form. (ii) The transform of the right-hand side of the integral equation is
By Exercise (10.3.3), Ff:(e) = [1 - F*(e)]/(J.Le), and(*) simplifies to m*(e) - (m* - m* F* F*)/(J.Le), which equals m*(e) since the quotient is 0 (by the renewal equation). Using the key renewal theorem, as t ~ oo,
lo [1t
0
1
looo [1- FE(u)]du = --JE(X2) a2 + JL2 = 2
FE(t -x)]dm(x) ~JL 0
2JL
by Exercise (10.3.3b). Therefore,
12. (i) Conditioning on X 1, we obtain
379
2JL
[10.6.13]-[10.6.16]
Solutions
Renewals
Thereforemd* = Fd* +m*Fd*. Alsom* = F* +m*F*,sothat
whence md* = Fd* + md* F*, the transform of the given integral equation. (ii)ArguingasinProblem(10.6.2), vd* = Fd*+2m*Fd*+v*Fd*wherev* = F*(l+2m*)/(l-F*) is the corresponding object in the ordinary renewal process. We eliminate v* to find that
by(*). Now invert.
13. Taking into account the structure of the process, it suffices to deal with the case I = 1. Refer to Example (10.4.22) for the basic notation and analysis. It is easily seen that fJ = (v - 1)A.. Now F(t) = 1- e-vA.t. Solve the renewal equation (10.4.24) to obtain
+lot
g(t) = h(t)
h(t- x) diii(x)
where iii(x) = vA.x is the renewal function associated with the interarrival time distribution Therefore g(t) = 1, andm(t) = ef3t.
F.
14. We have from Lemma (10.4.5) that p* = 1 - Fz + p* F*, where F* = Fy Fz. Solve to obtain
* p
=
1- Fz 1- F*F* y z
15. The first locked period begins at the time of arrival of the first particle. Since all future events may be timed relative to this arrival time, we may take this time to be 0. We shall therefore assume that a particle arrives at 0; call this the Oth particle, with locking time Yo. We shall condition on the time X 1 of the arrival of the next particle. Now lP'(L > t
I X1
= u) = {
lP'(Yo>t)
ifu>t,
lP'(Yo > u)lP'(L' > t - u)
ifu ~ t,
where L' has the same distribution as L; the second part is a consequence of the fact that the process 'restarts' at each arrival. Therefore lP'(L > t) = (1- G(t))lP'(X1 > t)
+lot
lP'(L > t - u)(1- G(u))fx 1 (u)du,
the required integral equation. If G(x) = 1 - e-11-x, the solution is lP'(L > t) = e-11-t, so that L has the same distribution as the locking times of individual particles. This striking fact may be attributed to the lack-of-memory property of the exponential distribution.
16. (a) It is clear that M(tp) is a renewal process whose interarrival times are distributed as X1 + Xz + · · · + XR where lP'(R = r) = pqr- 1 for r:::: 1. It follows that M(t) is a renewal process whose first interarrival time X(p) = inf{t : M(t) = 1} = p inf{t : M(tp) = 1}
380
Solutions [10.6.17]-[10.6.19]
Problems
has distribution function 00
lP'(X(p)::::: x) = LlP'(R = r)Fr(xjp). r=1
(b) The characteristic function QJp of Fp is given by QJp(t)
=
f
pqr-1100 eixt dFr(tfp)
r=1
=
-oo
f
pqr-14J(pt{
=
pqy(pt)
1 - qqy(pt)
r=1
where qy is the characteristic function of F. Now qy (pt) = 1 + i Jlpt + o(p) as p ,), 0, so that QJ (t)
=
P
1 + ijlpt + o(p) 1-itLt+o(1)
=
1 + o(l) 1-itLt
as p ,), 0. The limit is the characteristic function of the exponential distribution with mean fL, and the continuity theorem tells us that the process M converges in distribution as p ,), 0 to a Poisson process with intensity 1I fL (in the sense that the interarrival time distribution converges to the appropriate limit). (c) If M and N have the same fdds, then c/Jp(t) = qy(t), which implies that qy(pt) = qy(t)f(p + qqy(t)). Hence 1/f(t) = c/J(t)- 1 satisfies 1/f(pt) = q + pl/f(t) fort E JR. Now 1/f is continuous, and it follows as in the solution toProblem(5.12.15) that 1/f has the form 1/f(t) = 1+,Bt, implyingthatqy(t) = (1 +.Bt)- 1 for some .B E «:::. The only characteristic function of this form is that of an exponential distribution, and the claim follows.
17. (a) Let N(t) be the number of times the sequence has been typed up to the tth keystroke. Then N is a renewal process whose interarrival times have the required mean fL; we have that JE(N(t))/t --+ fL- 1 as t --+ oo. Now each epoch of time marks the completion of such a sequence with probability ( 1 0 ) 14 , so that
6
~lE(N(t)) = ~ t
t
t (-1-) (-1-) 14--+
n= 14
100
100
14
as t--+ oo,
implying that fL = 1028 . The problem with 'omo' is 'omomo' (i.e., appearances may overlap). Let us call an epoch of time a 'renewal point' if it marks the completion of the word 'omo', disjoint from the words completed at previous renewal points. In each appearance of 'omo', either the first 'o' or the second 'o' (but not both) is a renewal point. Therefore the probability un, that n is a renewal point, satisfies ( 1 0 ) 3 = un + un-2( 1 0 ) 2 . Average this over n to obtain
6
6
C~or = n~moo ~ j; {Un +un-2 C~or} = ~ + ~ C~or, and therefore fL = 106 + 102 . (b) (i) Arguing as for 'omo', we obtain p 3 = un + pun-1 + p 2un-2, whence p 3 = (1 + p + p 2 )/ fL. (if) Similarly, p 2q = Un + pqun-2, so that fL = (1 + pq)f(p 2q). 18. The fdds of {N(u) - N(t) : u 2:: t} depend on the distributions of E(t) and of the interarrival times. In a stationary renewal process, the distribution of E(t) does not depend on the value oft, whence {N(u) - N(t) : u 2:: t} has the same fdds as {N(u) : u 2:: 0}, implying that X is strongly stationary.
19. We use the renewal-reward theorem. The mean time between expeditions is B fL, and this is the mean length of a cycle of the process. The mean cost of keeping the bears during a cycle is B(B - 1)cfL, whence the long-run average cost is {d + B(B - 1)ctL/2}/(B fL).
i
381
11
Queues
11.2 Solutions. M/M/1 1. The stationary distribution satisfies 1f = equations
1f P
when it exists, where P is the transition matrix. The Plfn-1 lfn = - 1+p
lfn+1 +1+p
for n > 2, -
with l::~o :rri = 1, have the given solution. If p ;::: 1, no such solution exists. It is slightly shorter to use the fact that such a walk is reversible in equilibrium, from which it follows that 1f satisfies for n ;::: 1. 2. (i) This continuous-time walk is a Markov chain with generator given by g01 = Bo, gn,n+ 1 = Bnp/(1 + p) and gn,n-1 = Bn/(1 + p) for n ;::: 1, other off-diagonal terms being 0. Such a process is reversible in equilibrium (see Problem (6.15.16)), and its stationary distribution v must satisfy Vngn,n+1 = Vn+1gn+1,n· These equations may be written as for n ;::: 1. These are identical to the equations labelled(*) in the previous solution, with :rrn replaced by vnBn. It follows that Vn = C:rrn/Bn for some positive constant C. (ii) If Bo = A., Bn =A.+ f.-i for n ;::: 1, we have that
whence C = 2A. and the result follows. 3. Let Q be the number of people ahead of the arriving customer at the time of his arrival. Using the lack -of-memory property of the exponential distribution, the customer in service has residual servicetime with the exponential distribution, parameter f.-t, whence W may be expressed as S 1 + S2 + · · · + S Q, the sum of independent exponential variables, parameter f.-t. The characteristic function of W is
I Q)}
=E{ (f.-t~it)Q}
-c--_1_-,...,---P--,- = (1 _ p) 1-pf.-t/({t-it)
382
+p
(
1-i - A. ) . {t-A.-it
Solutions
M/M/1
[11.2.4]-[11.2.7]
This is the characteristic function of the given distribution. The atom at 0 corresponds to the possibility that Q = 0. 4. We prove this by induction on the value of i + j. If i + j = 0 then i = j = 0, and it is easy to check that rr(O; 0, 0) = 1 and A(O; 0, 0) = 1, A(n; 0, 0) = 0 for n ::: 1. Suppose then that K ::: 1, and that the claim is valid for all pairs (i, j) satisfying i + j = K. Let i and j satisfy i + j = K + 1. The last ball picked has probability i /(i + j) of being red; conditioning on the colour of the last ball, we have that rr(n;i,j)= "+i .rr(n-1;i-1,j)+ "+j .rr(n+1;i,j-1). l 1 l 1 Now (i - 1) + j = K = i + ( j - 1). Applying the induction hypothesis, we find that rr(n; i, j) = i
~j
{ A(n- 1; i - 1, j)- A(n; i - 1, j)}
+ i
~j
{ A(n + 1; i, j - 1)- A(n + 2; i, j - 1) }·
Substitute to obtain the required answer, after a little cancellation and collection of terms. Can you see a more natural way?
5.
Let A and B be independent Poisson process with intensities J.... and p, respectively. These processes generate a queue-process as follows. At each arrival time of A, a customer arrives in the shop. At each arrival-time of B, the customer being served completes his service and leaves; if the queue is empty at this moment, then nothing happens. It is not difficult to see that this queue-process is M(J....)/M(p,)/1. Suppose that A(t) = i and B(t) = j. During the time-interval [0, t], the order of arrivals and departures follows the schedule of Exercise (11.2.4), arrivals being marked as red balls and departures as lemon balls. The imbedded chain has the same distributions as the random walk of that exercise, and it follows that JP>(Q(t) = n I A(t) = i, B(t) = j) = rr(n; i, j). Therefore Pn (t) = 'E.i,j rr(n; i, j)JP>(A(t) = i)JP>(B(t) = j). 6.
With p
= A/ p,, the stationary distribution of the imbedded chain is, as in Exercise (11.2.1 ),
{ io - p) Jrn = iO - p2)pn-1 ~
if n = 0, if n::: 1.
In the usual notation of continuous-time Markov chains, go= by Exercise (6.10.11), there exists a constant c such that rro
Now'£.; rr;
=
;J.... (1- p),
Jrn
=
J....
and gn =
J....
+ p, for n :::
1, whence,
2 1 for n::: 1. 2 (/...: p,) (1- p )pn-
= 1, and therefore c = 2/... and Jrn = (1 -
p)pn as required. The working is reversible.
7. (a) Let Q; (t) be the number of people in the ith queue at timet, including any currently in service. The process Q1 is reversible in equilibrium, and departures in the original process correspond to arrivals in the reversed process. It follows that the departure process of the first queue is a Poisson process with intensity J...., and that the departure process of Q 1 is independent of the current value of Q1· (b) We have from part (a) that, for any given t, the random variables Q1 (t), Q2(t) are independent. Consider an arriving customer when the queues are in equilibrium, and let W; be his waiting time (before service) in the ith queue. With T the time of arrival, and recalling Exercise (11.2.3), JP>(W1
= 0,
W2
= 0)
> JP>(Q;(T)
= (1 -
= 0 fori= 1, 2) = lP'(Q1 (T) = O)JP>(Q2(T) = 0) = lP'(W1 = O)JP>(W2 = 0).
P1)(1 - P2)
Therefore W1 and W2 are not independent. There is a slight complication arising from the fact that T is a random variable. However, T is independent of everybody who has gone before, and in particular of the earlier values of the queue processes Q;.
383
[11.3.1]-[11.4.1]
Solutions
Queues
11.3 Solutions. M/G/1 1.
In equilibrium, the queue-length Qn just after the nth departure satisfies
where An is the number of arrivals during the (n + l)th service period, and h(m) = 1 - 8mo· ·Now Qn and Qn+l have the same distribution. Take expectations to obtain
0 = lE(An) - IP'(Qn > 0), where lE (An) = A.d, the mean number of arrivals in an interval of length d. Next, square (*) and take expectations:
Use the facts that An is independent of Qn, and that Qnh(Qn) 0
= {(A.d) 2 + A.d} + IP'(Qn
> 0)
=
Qn, to find that
+ 2{ (A.d- l)lE(Qn)- A.dlP'(Qn >
0)}
and therefore, by (** ), 2p- p2 lE(Qn) = 2 (1 _ p)
From the standard theory, Ms satisfies Ms(s) = Ms(s - A. + A.Ms(s)), where Ms(e) ~-t/(J-t- e). Substitute to find that x = Ms (s) is a root of the quadratic A.x 2 - x(A. + 1-t- s) + 1-t = 0. For some small positive s, M B (s) is smooth and non-decreasing. Therefore M B (s) is the root given.
2.
3. Let Tn be the instant of time at which the server is freed for the nth time. By the lack-of-memory property of the exponential distribution, the time of the first arrival after Tn is independent of all arrivals prior to Tn, whence Tn is a 'regeneration point' of the queue (so to say). It follows that the times which elapse between such regeneration points are independent, and it is easily seen that they have the same distribution.
11.4 Solutions. G/M/1 1. The transition matrix of the imbedded chain obtained by observing queue-lengths just before arrivals is 0 ... 0 ...
ao
)
The equation :rr = :rr P A may be written as 00
Trn
= L ai:rrn+i-1
for n 2:: 1.
i=O
It is easily seen, by adding, that the first equation is a consequence of the remaining equations, taken in conjunction with L:Q" :rri = 1. Therefore :rr is specified by the equation for :rrn, n 2:: 1.
384
Solutions [11.4.2]-[11.5.2]
G/G/1
The indicated substitution gives 00
en= en-1 l.:a;ei i=O
which is satisfied whenever e satisfies
It is easily seen that A(8) = Mx(~-t(8 - I)) is convex and non-decreasing on [0, 1], and satisfies A(O) > 0, A(l) = 1. Now A'(l) = ~-tlE(X) = p- 1 > 1, implying that there is a unique 'f/ E (0, 1) such that A('f/) = 'f'J. With this value of 'f'J, the vector 1r given by Trj = (1 - TJ)'f/1, j ;:: 0, is a stationary distribution of the imbedded chain. This 1r is the unique such distribution because the chain is irreducible.
2. (i) The equilibrium distribution is Trn = (1- TJ)'f'Jn for n 2: 0, with mean l:~o nnn = TJ/0- TJ). (ii) Using the lack-of-memory property of the service time in progress at the time of the arrival, we see that the waiting time may be expressed as W = S1 + S2 + · · · + S Q where Q has distribution 1r, given above, and the Sn are service times independent of Q. Therefore JE(W)
= lE(S1)lE(Q) = ___!}_j£__. (1- 'f/)
We have that Q(n+) = 1 + Q(n-) a.s. foreachintegern, whencelim 1 _, 00 lP'(Q(t) = m) cannot exist. Since the traffic intensity is less than 1, the imbedded chain is ergodic with stationary distribution as in Exercise (11. 4.1).
3.
11.5 Solutions. G/G/1 1. Let Tn be the starting time of the nth busy period. Then Tn is an arrival time, and also the beginning of a service period. Conditional on the value of Tn, the future evolution of the queue is independent of the past, whence the random variables {Tn+l - Tn : n :::: 1} are independent. It is easily seen that they are identically distributed. 2. If the server is freed at time T, the time I until the next arrival has the exponential distribution with parameter J-t (since arrivals form a Poisson process). By the duality theory of queues, the waiting time in question has moment generating function Mw(s) = (1- 0/(1- M1(s)) where M1(s) = ~-t/(J-t- s) and = lP'(W > 0). Therefore,
s
s
Mw(s)=
s~-t( 1 -0 +(1-S),
~-tO-n-s
the moment generating function of a mixture of an atom at 0 and an exponential distribution with parameter ~-t(1 - S). If G is the probability generating function of the equilibrium queue-length, then, using the lackof-memory property of the exponential distribution, we have that Mw(s) = G(~-t/(J-t- s)), since W is the sum of the (residual) service times of the customers already present. Set u = ~-t/(J-t- s) to find that G(u) = (1- 0/(1- su), the generating function of the mass function f(k) = (1- Osk fork::: 0.
385
[11.5.3]-(11.7.2]
Solutions
Queues
It may of course be shown that ~ is the smallest positive root of the equation x = M x (~-t(x where X is a typical interarrival time. 3.
1)),
We have that 1- G(y)
= lP'(S- X>
y)
=
lx:;
lP'(S > u
+ y)dFx(u),
y
E
JR,
where Sand X are typical (independent) service and interarrival times. Hence, formally, dG(y) = - {
lo
00
dlP'(S > u
+ y) dFx(u)
= dy
1-y ~-te-JJ,(u+y) 00
dFx(u),
since fs(u + y) = e-JJ,(u+y) if u > -y, and is 0 otherwise. With F as given,
1
x F(x- y)dG(y) = {{ {I -oo }}-oo
-7Je-JJ,(l-1})(x-y)}~-te-JJ,(u+y) dFx(u)dy.
First integrate over y, then over u (noting that F x (u) = 0 for u < 0), and the double integral collapses to F(x), when x ::::: 0.
11.6 Solution. Heavy traffic 1.
Qp has characteristic function
Therefore the characteristic function of (1- p)Qp satisfies 1- p 1 cl>p((l- p)t) = 1- pei(I-p)t --+ 1- it
asptl.
The limit characteristic function is that of the exponential distribution, and the result follows by the continuity theorem.
11.7 Solutions. Networks of queues 1. The first observation follows as in Example (11.7.4). The equilibrium distribution is given as in Theorem (11.7.14) by c
n(n) =
ni -a·
a. e II -'n·'
'
1 --
i=I
forn = (nJ,nz, .. . ,nc) E 71/,
z·
the product of Poisson distributions. This is related to Bartlett's theorem (see Problem (8.7.6)) by defining the state A as 'being in station i at some given time'.
2. The number of customers in the queue is a birth-death process, and is therefore reversible in equilibrium. The claims follow in the same manner as was argued in the solution to Exercise (11.2.7).
386
Solutions
Problems
[11.7.3]-[11.8.1]
3. (a) We may take as state space the set {0, 11, 111 , 2, 3, ... }, where i E {0, 2, 3, ... } is the state of having i people in the system including any currently in service, and 11 (respectively 111 ) is the state of having exactly one person in the system, this person being served by the first (respectively second) server. It is straightforward to check that this process is reversible in equilibrium, whence the departure process is as stated, by the argument used in Exercise (11.2.7). (b) This time, we take as state space the set {0', 0 11 , 11 , 1", 2, 3, ... } having the same states as in part (a) with the difference that O' (respectively O") is the state in which there are no customers present and the first (respectively second) server has been free for the shorter time. It is easily seen that transition from 01 to 111 has strictly positive probability whereas transition from 111 to 0 1 has zero probability, implying that the process is not reversible. By drawing a diagram of the state space, or otherwise, it may be seen that the time-reversal of the process has the same structure as the original, with the unique change that states 01 are 011 are interchanged. Since departures in the original process correspond to arrivals in the time-reversal, the required properties follow in the same manner as in Exercise (11.2.7). 4. The total time spent by a given customer in service may be expressed as the sum of geometrically distributed number of exponential random variables, and this is easily shown to be exponential with parameter 8~-t. The queue is therefore in effect aM(A.)/M(8~-t)ll system, and the stationary distribution is the geometric distribution with parameter p = A./(8~-t), provided p < 1. As in Exercise (11.2.7), the process of departures is Poisson. Assume that rejoining customers go to the end of the queue, and note that the number of customers present constitutes a Markov chain. However, the composite process of arrivals is not Poisson, since increments are no longer independent. This may be seen as follows. In equilibrium, the probability of an arrival of either kind during the time interval (t, t +h) is A.h + P~-tO- 8)h +o(h) = (A.j8)h + o(h). If there were an arrival of either kind during (t- h, t), then (with conditional probability 1 - O(h)) the queue is non-empty at time t, whence the conditional probability of an arrival of either kind during (t, t +h) is A.h + ~-tO - 8)h + o(h); this is of a larger order of magnitude than the earlier probability (A./8)h + o(h). 5. For stations r, s, we write r --+ s if an individual at r visits s at a later time with a strictly positive probability. Let C comprise the station j together with all stations i such that i --+ j. The process restricted to C is an open migration process in equilibrium. By Theorem (11. 7 .19), the restricted process is reversible, whence the process of departures from C via j is a Poisson process with some intensity t;. Individuals departing C via j proceed directly to k with probability Ajk~j(nj)
Ajk
1-tj~j(nj)+l:.rrtCAjr~j(nj) =
1-tj+l:.rrtCAjr'
independently of the number nj of individuals currently at j. Such a thinned Poisson process is a Poisson process also (cf. Exercise (6.8.2)), and the claim follows.
11.8 Solutions to problems 1. Although the two cases may be done together, we choose to do them separately. When k = l, the equilibrium distribution 1r satisfies: J-t7r1 - A.1ro = 0, J-t1rn+1 -(A.+ J-t)7rn + A'lrn-1 = 0,
1::::; n < N,
-J-t'lrN + A'lrN-1 = 0,
a system of equations with solution 7rn
= 7ro(A./ J-t)n for 0::::; n
N 'Jf-1 ="'\:'(A./ )n =
o
L n=O
1-t
387
1
-
::::; N, where (if A.
(A./ )N+1 1-t
1-(A/~-t)
=f. J-t)
[11.8.2]-[11.8.3]
Solutions
Queues
Now let k = 2. The queue is a birth-death process with rates A.·- { 1 -
A. 0
ifi < N, ifi 2:: N,
/Li =
{
/L 2JL
ifi = 1, ifi 2:: 2.
It is reversible in equilibrium, and its stationary distribution satisfies A.ini that ni = 2pino for 1 :::: i ::=: N, where p = A./(2/L) and
=
/Li+lni+l· We deduce
N
no'= 1 + 2.:2pi. i=l
2. The answer is obtainable in either case by following the usual method. It is shorter to use the fact that such processes are reversible in equilibrium. (a) The stationary distribution1r satisfies nnA.p(n) = nn+l/L for n 2:: 0, whence nn = nopn fn! where p = A./JL. Therefore nn = pne-P fn!. (b) Similarly, n-1 n 2::0, nn = nop n P m = nop n 2 - ~n(n-1) ",
II ( )
m=O
where oo no'= LPn(~)Zn(n-1). I
n=O
At the instant of arrival of a potential customer, the probability q that she joins the queue is obtained by conditioning on its length: oo oo oo 1 =2.: p(n)nn =no 2.: pn2-n-zn
q
n=O
I
n=O
n=O
P
3. First method. Let (Ql, Qz) be the queue-lengths, and suppose they are in equilibrium. Since Q, is a birth-death process, it is reversible, and we write (h(t) = Q,(-t). The sample paths of Q, have increasing jumps of size 1 at times of a Poisson process with intensity A.; these jumps mark arrivals at the cash desk. By reversibility, Q, has the same property; such increasing jumps for Q, are decreasing jumps for Q 1, and therefore the times of departures from the cash desk form a Poisson process with intensity A.. Using the same argument, the quantity Q, (t) together with the departures prior tot have the same joint distribution as the quantity Q, (-t) together with all arrivals after -t. However Q, (-t) is independent of its subsequent arrivals, and therefore Q, (t) is independent of its earlier departures. It follows that arrivals at the second desk are in the manner of a Poisson process with intensity A., and that Qz(t) is independent of Q, (t). Departures from the second desk form a Poisson process also. Hence, in equilibrium, Q, is M(A.)/M(J.LI)/1 and Qz is M(A.)/M(J.Lz)/1, and they are independent at any given time. Therefore their joint stationary distribution is nmn =IP'(Q,(t) =m, Qz(t) =n) = (1-p,)(l-pz)PiP!J.
= A./ J.Li. Second method. The pair (Ql (t), Qz(t)) is a bivariate Markov chain. A stationary distribution (nmn : m, n 2:: 0) satisfies where Pi
(A.+ ILl+ J.Lz)nmn
= A.nm-l,n +
J.Linm+l,n-1 + J.Lznm,n+l•
388
m,n 2::1,
Solutions [11.8.4]-[11.8.5]
Problems
together with other equations when m = 0 or n = 0. It is easily checked that these equations have the solution given above, when Pi < 1 for i = 1, 2.
4. Let Dn be the time of then th departure, and let Q n = Q ( Dn +) be the number of waiting customers immediately after Dn. We have in the usual way that Qn+l = An+ Qn - h(Qn), where An is the number of arrivals during the (n + l)th service time, and h(x) = min{x, m}. Let G(s) = l:~o :rr;si be the equilibrium probability generating function of the Qn. Then, since Qn is independent of An,
where
lE(sAn)
=
fooo
e)..u(s-l) fs(u)
du
= Ms (A.(s
-
1)),
M s being the moment generating function of a service time, and
Combining these relations, we obtain that G satisfies
smG(s) = Ms(A.(s-
1)) { G(s) +
fcsm - si):rr; }• 1=0
whenever it exists. Finally suppose that m = 2 and Ms(e) = M/(M- 8). In this case, M{rro(s + 1) + :rr1s} G(s) = '------'----=-----=----'M(s + 1)- A.s 2
Now G(l) = 1, whence M(2:rro + :rr1) = 2M- A.; this implies in particular that 2M- A. > 0. Also G(s) converges for Is I _:::: 1. Therefore any zero of the denominator in the interval [-1, 1] is also a zero of the numerator. There exists exactly one such zero, since the denominator is a quadratic which takes the value -A. at s = -1 and the value 2M - A. at s = 1. The zero in question is at
and it follows that rro + (:rro + :rr1 )so
= 0.
Solving for rro and :rr1, we obtain
1-a -as
G(s) = -1 - ,
where a= 2A./{M +
5.
J M2 + 4A.J-[}.
Recalling standard M/G/1 theory, the moment generating function M B satisfies
Ms(s) = Ms(s- A.+ A.Ms(s)) = -----,----M------,M-{s-A.+A.Ms(s)}
whence M B (s) is one of (A.+ M- s)
± V(A. + 2)..
389
M- s) 2 - 4A.J-[
[11.8.6]-[11.8. 7]
Solutions
Queues
Now M B (s) is non-decreasing in s, and therefore it is the value with the minus sign. The density function of B may be found by inverting the moment generating function; see Feller (1971, p. 482), who has also an alternative derivation of M B. As for the mean and variance, either differentiate MB, or differentiate(*). Following the latter route, we obtain the following relations involving M (= MB): 2AM M 1 + M
+ (s -A -
JL)M 1
= 0,
2AM M 11 + 2A(M 1) 2 + 2M 1 + (s -A - JL)M 11 = 0.
Sets = 0 to obtain M 1 (0) = (JL -A)- 1 and M 11 (0) = 2J.L(J.L-A)- 3 , whence the claims are immediate.
6. (i) This question is closely related to Exercise (11.3.1). With the same notation as in that solution, we have that
where h(x) = min{l, x}. Taking expectations, we obtain IP'(Qn > 0) = JE(An) where lE(An) =
lX>
lE(An IS= s)dFs(s) = AlE(S) = p,
and S is a typical service time. Square (*) and take expectations to obtain p(l- 2p)
lE(Qn) =
+ JE(A~+l)
2 (1 _ p)
,
where JE(A~) is found (as above) to equal p +A 2 JE(S 2 ). (ii) If a customer waits for time W and is served for time S, he leaves behind him a queue-length which is Poisson with parameter A(W + S). In equilibrium, its mean satisfies AlE(W + S) = lE(Qn), whence lE(W) is given as claimed. (iii) JE(W) is a minimum when JE(S 2 ) is minimized, which occurs when Sis concentrated at its mean. Deterministic service times minimize mean waiting time. 7. Condition on arrivals in (t, t+h). If there are no arrivals, then Wt+h ::::= x if and only ifW1 ::::= x+h. If there is an arrival, and his service time is S, then Wt+h ::::= x if and only if W1 ::::= x + h- S. Therefore F(x; t +h)= (l - Ah)F(x + h; t) + Ah
Jor+h F(x + h- s; t) dFs(s) + o(h).
Subtract F(x; t), divide by h, and take the limit ash ,).. 0, to obtain the differential equation. We take Laplace-Stieltjes transforms. Integrating by parts, for e :::: 0, {
lco,oo) {
lco,oo) {
lco,oo)
eex dh(x) = -h(O)- B{Mu(B) - H(O)},
eex dH(x) = Mu(B)- H(O),
eex dlP'(U
+ S:::: x) =
Mu(B)Ms(B),
and therefore 0 = -h(O)- B{Mu(B)- H(O)}
+ AH(O) + AMu(B){Ms(B)- 1}.
390
Solutions [11.8.8]-[11.8.10]
Problems
Set e = 0 to obtain that h(O) = 'AH(O), and therefore 1 H(O) = -eMu(e){'A(Ms(e)-
1)- e}.
Take the limit as e --.. 0, using L'Hopital's rule, to obtain H(O) = 1- 'AJE.(S) = 1 - p. The moment generating function of U is given accordingly. Note that Mu is the same as the moment generating function of the equilibrium distribution of actual waiting time. That is to say, virtual and actual waiting times have the same equilibrium distributions in this case.
8. In this case U takes the values 1 and -2 each with probability ~ (as usual, U = S- X where S and X are typical (independent) service and interarrival times). The integral equation for the limiting waiting time distribution function F becomes F(O) = ~ F(2),
F(x)
=
HF(x- 1) + F(x + 2)}
The auxiliary equation is e3 - 28 + 1 = 0, with roots 1 and- ~(1 can contribute, whence F(x) =A+ B (
forx = 1,2, ....
± J5).
Only roots lying in [-1, 1]
-l: JSr
for some constants A and B. Now F(x) -->- 1 as x -->- oo, since the queue is stable, and therefore A= 1. Using the equation for F(O), we find that B = ~(1-
JS).
9. Q is a M('A)/M(JL)/oo queue, otherwise known as an immigration-death process (see Exercise (6.11.3) and Problem (6.15.18)). As found in (6.15.18), Q(t) has probability generating function G(s, t) = { 1 + (s- l)e-JLt} I exp{p(s- 1)(1 - e-JLt)}
where p ='A/ IL· Hence JE.(Q(t)) = Ie-JLt
+ p(l- e-JLt),
lP'(Q(t) =0) = (1-e-JLti exp{-p(1-e-ILt)}, as t -->- oo. If JE.(J) and JE.(B) denote the mean lengths of an idle period and a busy period in equilibrium, we have that the proportion of time spent idle is JE.(/)/{lE.(J) + JE.(B)}. This equals limt-H>o lP'(Q(t) = 0) = e-P. Now JE.(J) = A- 1, by the lack-of-memory property of the arrival process, so that JE.(B) = (eP- 1)/'A.
10. We have in the usual way that Q(t
+ 1) =At+ Q(t)- min{l, Q(t)}
where At has the Poisson distribution with parameter A. When the queue is in equilibrium, JE.(Q(t)) = + 1)), and hence
JE.(Q(t
lP'(Q(t) > 0) = JE.(min{l, Q(t)}) = JE.(At) ='A. We have from (*)that the probability generating function G(s) of the equilibrium distribution of Q(t) (= Q) is
G(s) = JE.(sAt)JE.(sQ-min{1,Q}) = e).(s-1){lE.(sQ-1 l(Q:::l))
391
+ lP'(Q =
0) }.
[11.8.11]-[11.8.13]
Solutions
Queues
Also, G(s) = JE.(sQ I{Q:::1J) + JP>(Q = 0), and hence G(s) = whence
e).(s- 1) {
~G(s) +
( 1-
~) (1- A)}
(1 - s)(l -A) G(s) = 1- se -).(s -1).
The mean queue length is G' (1) = ~A(2 - A)/(1 -A). Since service times are of unit length, and arrivals form a Poisson process, the mean residual service time of the customer in service at an arrival time is ~, so long as the queue is non-empty. Hence A
JE.(W) = JE.(Q)- ~JP>(Q > 0) = - - 2(1- A)
11. The length B of a typical busy period has moment generating function satisfying M B (s) = exp{s- A+ AMs(s)}; this fact may be deduced from the standard theory ofM/G/1, or alternatively by a random-walk approach. Now T may be expressed as T = I + B where I is the length of the first idle period, a random variable with the exponential distribution, parameter A. It follows that MT(s) = AMs(s)/(A- s). Therefore, as required, (A- s)MT(s) = Aexp{s- A+ (A- s)MT(s) }. If A :::: 1, the queue~length at moments of departure is either null persistent or transient, and it follows that JE.(T) = oo. If A < 1, we differentiate(*) and sets = 0 to obtain AlE.(T)- 1 =A 2 JE.(T), whence JE.(T) = {A(l - A)} -1. 12. (a) Q is a birth-death process with parameters Ai = A, /-[i = {[, and is therefore reversible in equilibrium; see Problems (6.15.16) and (11.8.3). (b) The equilibrium distribution satisfies krci = {[7ri+1 fori :::: 0, whence Jri = (1- p)pi where p = A/ f.-[. A typical waiting time W is the sum of Q independent service times, so that
Mw(s)=GQ(Ms(s))=
1-p
1- P{[/({[- s)
=
(1-p)(f.-[-S) . f.-[(1- p)- s
(c) See the solution to Problem (11.8.3). (d) Follow the solution to Problem (11.8.3) (either method) to find that, at any timet in equilibrium, the queue lengths are independent, the jth having the equilibrium distribution ofM(A)IM({[j )/1. The joint mass function is therefore
where
Pj
=A/{[J.
13. The size of the queue is a birth-death process with rates Ai =A, /-[i = {[ min{i, k}. Either solve the equilibrium equations in order to find a stationary distribution 7r, or argue as follows. The process is reversible in equilibrium (see Problem (6.15.16)), and therefore Ailri = /-[i+17ri+1 for all i. These 'balance equations' become ifO~i
ifi:::: k.
392
Solutions [11.8.14]-[11.8.14]
Problems
These are easily solved iteratively to obtain
Tri
=
{
noai I i! no(aj k)i kk I k!
ifO:Si ::Sk, ifi:::: k
where a = A./ JL. Therefore there exists a stationary distribution if and only if J.. . < kJL, and it is given accordingly, with
The cost of having k servers is oo (a/ k)i kk Ck = Ak + Bno L)- k + 1)-'---'--::-:-i=k k!
where no = no(k). One finds, after a little computation, that
c1
Ba =A+, 1-a
2Ba 2 4-a
C2 =2A+ - 2.
Therefore
a 3 (A - B)+ a 2 (2B - A) - 4a(A +B)+ 4A C2- Cl =
(1 - a)(4- a2)
Viewed as a function of a, the numerator is a cubic taking the value 4A at a= 0 and the value -3B at a= l. This cubic has a unique zero at some a* E (0, 1), and C1 < C2 if and only ifO
14. The state of the system is the number Q(t) of customers within it at timet. The state 1 may be divided into two sub-states, being a1 and a2, where O"i is the state in which server i is occupied but the other server is not. The state space is therefore S = {0, a1, a2, 2, 3, ... }. The usual way of finding the stationary distribution, when it exists, is to solve the equilibrium equations. An alternative is to argue as follows. If there exists a stationary distribution, then the process is reversible in equilibrium if and only if
for all sequences i1, i2, ... , ik of states, where G = (guv)u,vES is the generator of the process (this may be shown in very much the same way as was the corresponding claim for discrete-time chains in Exercise (6.5.3); see also Problem (6.15.16)). It is clear that(*) is satisfied by this process for all sequences of states which do not include both a1 and a2; this holds since the terms guv are exactly those of a birth-death process in such a case. In order to see that (*) holds for a sequence containing both a1 and a2, it suffices to perform the following calculation: go,al gal ,2g2,azgaz,O = ( ~ J....)J....J.L2/Ll = go,az gaz,2g2,al gal ,0·
Since the process is reversible in equilibrium, the stationary distribution i= v. Therefore
1r:
satisfies Truguv
Trvgvu for all u, v E S, u
and hence Tra1
= -2
J.. .
ILl
no,
')...2
Tru = - - 2JL 1/L2
393
(
')...
JL1
+ JL2
)u-2 no
for u ::=: 2.
[11.8.15]-[11.8.17]
Solutions
Queues
This gives a stationary distribution if and only if).. < JL1 + f.i-2, under which assumption no is easily calculated. A similar analysis is valid if there are s servers and an arriving customer is equally likely to go to any free server, otherwise waiting in tum. This process also is reversible in equilibrium, and the stationary distribution is similar to that given above.
15. We have from the standard theory that QJL has as mass function Trj = (1 - IJ)IJj, j 0::: 0, where IJ is the smallest positive root of the equation x = eJL(x-I). The moment generating function of (1- f.L- 1)QJL is 1 -1] _ . MJL(8)=E(exp{8Cl-f.L-l)QJL})= 1 1 - IJeeO-JL l Writing f.L = 1 + E, we have by expanding eiLC'7-l) as a Taylor series that IJ = IJ(E) = 1- 2E + o(E) as E ,j, 0. This gives 2E + O(E) 2E + O(E) 2 = 1 - (1 - 2E)(l +BE)+ o(E) = (2- B)E + o(E) --+ 2- e
MJL(B)
as E ,j, 0, implying the result, by the continuity theorem.
16. The numbers P (of passengers) and T (of taxis) up to timet have the Poisson distribution with respective parameters nt and rt. The required probabilities Pn = IP'(P = T + n) have generating function 00
00
2.: 2.: IP'(P = m + n)IP'(T = m)zn n=-oo
n=-OOm=O 00
=
2.: IP'(T = m)z-mG p(z) m=O
= GT(Z-l)Gp(Z) = e-(rr+r)te(rrz+rz-l)t, in which the coefficient of zn is easily found to be that given.
17. Let N(t) be the number of machines which have arrived by timet. Given that N(t) = n, the times T1, T2, ... , Tn of their arrivals may be thought of as the order statistics of a family of independent uniform variables on [0, t], say U1, U2, ... , Un; see Theorem (6.12.7). The machine which arrived at time
ui
is, at time t, in the X -stage} { a(t) in the Y -stage with probability f3 ( t) 1 - a(t) - f3(t) repaired
where a(t) = IP'(U +X > t) and f3(t) = IP'(U +X :=: t < U +X+ Y), where U is uniform on [0, t], and (X, Y) is a typical repair pair, independent of U. Therefore IP'(U(t) = . V(t) = k I N(t) = n) = n! a(t)j f3(ti (1 - a(t) - f3(t))n-k- j
1"lkl(· · n 1·-k)'·
1·
implying that oo e-At (At)n
IP'(U(t) = j, V(t) = k) =
2.:
n= 0
n.1
.,
1·
394
IP'(U(t) = j, V(t) = k I N(t) = n)
k!
,
Solutions [11.8.18]-[11.8.19]
Problems
18. The maximum deficit Mn seen up to and including the time of the nth claim satisfies
Mn
= max{Mn-1• t ( K j - Xj)} = max{O, U1, U1 + U2, ... , U1 + U2 + · · · + Un}, ;=1
where the Xj are the inter-claim times, and Uj = Kj - Xj. We have as in the analysis of G/G/1 that Mn has the same distribution as Vn = max{O, Un, Un + Un-1, ... , Un + Un-1 + · · · + Ul}, whence Mn has the same distribution as the (n + l)th waiting time in a M(A.)/G/1 queue with service times Kj and interarrival times Xj. The result follows by Theorem (11.3 .16).
19. (a) Look for a solution to the detailed balance equations A.rri = (i + l)JL7ri+1, 0 that the stationary distribution is given by Jri = (pi/ i !)no. (b) Let Pc be the required fraction. We have by Little's theorem (10.5.18) that
Pc =
A.(rrc-1 - Jrc)
= p(rrc-1 - rrc),
~
i < s, to find
c :=:: 2,
fL
and P1 = rr1, where Jrs is the probability that channels 1, 2, ... , s are busy in a queue M/M/s having the property that further calls are lost when all s servers are occupied.
395
12 Martingales
12.1 Solutions. Introduction 1.
(i) We have that E(Ym) = E{E(Ym+l I Tm)} = E(Ym+J), and the result follows by induction. (ii) For a submartingale, E(Ym) ::0 E{E(Ym+l I Tm)} = E(Ym+d, and the result for supermartingales follows similarly. 2.
We have that
if m :::: 1, since Tn s; Tn+m-1· Iterate to obtain E(Yn+m I Tn) = E(Yn I Tn) = Yn. 3.
(i) Znt-t-n has mean 1, and
where Tn = cr(ZJ, Zz, ... , Zn). (ii) Certainly 1'/Zn ::0 1, and therefore it has finite mean. Also,
where the Xi are independent family sizes with probability generating function G. Now G(ry) = ry, and the claim follows.
4.
(i) With Xn denoting the size of the nth jump,
where Tn = cr(X 1, Xz, ... , Xn). Also EISn I ::0 n, so that {Sn} is a martingale. (ii) Similarly E(S;) = var(Sn) = n, and
(iii) Suppose the walk starts at k, and there are absorbing barriers at 0 and N (:::: k). LetT be the time at which the walk is absorbed, and make the assumptions that E(ST) =So, E(Sf- T) = Then the probability Pk of ultimate ruin satisfies
sg.
0 · Pk
+N
· (1 - Pk) = k,
396
Solutions
Introduction
[12.1.5]-[12.1.9]
and therefore Pk = 1 - (k/ N) and lE(T) = k(N - k). 5.
(i) By Exercise (12.1.2), for r ::=: i, lE(YrYi)
= lE{lE(YrYi
I .rj)}
= lE{YilE(Yr
I .rj)}
= lE(Yh,
an answer which is independent of r. Therefore ifi_:::=j_:::=k.
(ii) We have that lE{ (Yk - YJ ) 2 l.rt} = lE(Yfl .rj) - 21E(Yk YJ I .rj)
+ lE(Y}I
.rj ).
NowlE(YkYj IJ=i) =lE{lE(YkYj I Jj)j J:'i} =E(Yf 1 J:'i),andthec1aimfo1lows. (iii) Taking expectations of the last conclusion,
j.:::: k. Now {lE(YJ) : n :;:: 1} is non-decreasing and bounded, and therefore converges. Therefore, by(*), { Yn : n :;:: 1} is Cauchy convergent in mean square, and therefore convergent in mean square, by Problem (7.11.11).
6.
(i) Using Jensen's inequality (Exercise (7.9.4)),
(ii) It suffices to note that lx I, x 2 , and x+ are convex functions of x; draw pictures if you are in doubt about these functions. 7. (i) This follows just as in Exercise (12.1.6), using the fact that u{lE(Yn+l I .7='n)} ::=: u(Yn) in this case. (ii) The function x+ is convex and non-decreasing. Finally, let {Sn : n ::=: 0} be a simple random walk whose steps are +1 with probability p (= 1 - q > ~) and -1 otherwise. If Sn < 0, then lE (ISn+III .7='n)
= p(ISnl-1) + q(ISnl + 1) =
ISnl- (p- q) < ISnl;
note that lP'(Sn < 0) > 0 if n ::=: 1. The same example suffices in the remaining case. 8.
Clearly lElA. -nlfr(Xn)l _::::A. -n sup{llfr(j)l : j
E
S}. Also,
lE(lfr(Xn+J) I .7='n) = LPXn,}lfr(j) _:::: A.lfr(Xn) }ES
where .7='n = cr(X 1, Xz, ... , Xn). Divide by A. n+l to obtain that the given sequence is a supermartingale. 9. Since var(ZJ) > 0, the function G, and hence also Gn, is a strictly increasing function on [0, 1]. Since 1 = Gn+l (Hn+l (s)) = Gn(G(Hn+l (s))) and Gn(Hn(s)) = 1, we have that G(Hn+l (s)) = Hn(s). With .7='m = cr(Zk : 0 _:::: k _:: : m), lE(Hn+l (s) 2 n+l I .7='n) = G(Hn+l (s))Zn = Hn(s)Zn.
397
[12.2.1]-[12.3.2]
Solutions
Martingales
12.2 Solutions. Martingale differences and Hoeffding's inequality 1. Let:Fi = a({Vj, Wj: 1 _::: j _::: i}) and Yi = lE(Z I :fl). With Z(j) the maximal worth attainable without using the jth object, we have that lE(Z(j)
I:FJ) =
lE(Z(j)
I0-d.
Z(j) _::: Z _::: Z(j)
+ M.
Take conditional expectations of the second inequality, given :F) and given :F) -I , and deduce that IYj - YJ-!1 _::: M. Therefore Y is a martingale with bounded differences, and Hoeffding's inequality yields the result. 2. Let :fi be the a-field generated by the (random) edges joining pairs (va, Vb) with 1 _::: a, b _::: i, and let Xi = lE(X I :fi ). We write x (j) for the minimal number of colours required in order to colour each vertex in the graph obtained by deleting Vj. The argument now follows that of the last exercise, using the fact that X (j) _::: X _::: X (j) + 1.
12.3 Solutions. Crossings and convergence 1.
Let T1
= min{n:
Yn
:=:::
b}, Tz
= min{n
Tzk-1 = min{n > Tzk-2 : Yn
> T1 : Yn _:::a}, and define Tk inductively by :=:::
Tzk = min{n > Tzk-1 : Yn _:::a}.
b},
The number of downcrossings by time n is Dn(a, b; Y) = max{k: Tzk _::: n}. (a) Between each pair of upcrossings of [a, b ], there must be a downcrossing, and vice versa. Hence IDn(a, b; Y)- Un(a, b; Y)l _::: 1. (b) Let Ii be the indicator function of the event that i
E
(Tzk-1, Tzk] for some k, and let
n
Zn =
L li(Yi- Yi-J),
n
:=:::
0.
i=l
It is easily seen that
Zn _::: -(b- a)Dn(a, b; Y)
+ (Yn- b)+,
whence
(b- a)lEDn(a, b; Y) _::: lE{ (Yn- b)+} -lE(Zn). Now Ii is :Fi-1-measurable, since
Ui
= 1} =
U({Tzk-1 ::: i- 1} \ {Tzk::: i -1}). k
Therefore,
since In :=::: 0 andY is a submartingale. It follows that lE(Zn) :=::: lE(Zn_J) :=::: · · · :=::: lE(Zo) = 0, and the final inequality follows from ( *). 2. If Y is a supermartingale, then - Y is a submartingale. Upcrossings of [a, b] by Y correspond to downcrossings of [- b, -a] by - Y, so that lEUn (a, b·, Y) =lEDn (-b , - a,· -Y) < -
398
lE{(- Y +a)+} n ba
JE{(Y -a)-} = __bn_ __ a ,
Solutions
Stopping times
[12.3.3]-[12.4.5]
by Exercise (12.3.1). If a, Yn 2':: 0 then (Yn -a)- :::::a. 3. The random sequence {1/r(Xn) : n :=:: 1} is a bounded supermartingale, which converges a.s. to some limit Y. The chain is irreducible and persistent, so that each state is visited infinitely often a.s.; it follows that limn-+oo 1/r(Xn) cannot exist (a.s.) unless 1fr is a constant function.
I:i"'
4. Y is a martingale since Yn is the sum of independent variables with zero means. Also IP'(Zn =f. 0) = L:j"' n- 2 < oo, implying by the Borel-Cantelli lemma that Zn = 0 except for finitely many values of n (a.s.); therefore the partial sum Yn converges a.s. as n -+ oo to some finite limit. It is easily seen that an = San -1 and therefore an = 8 · sn- 2 , if n :=:: 3. It follows that IYn I :=:: ian if and only if IZn I = an. Therefore
which tends to infinity as n -+ oo.
12.4 Solutions. Stopping times 1.
We have that n
U ({T1=k}n{T2=n-k}), k=O { max{T1, T2}::::: n} = m : : : n} n {T2::::: n}, { min{T1, T2}::::: n} = m : : : n} U {T2::::: n}. {T1+T2=n}=
Each event on the right-hand side lies in :Fn.
2.
Let :Fn = a(X1, X2, ... , Xn) and Sn = X1 + X2 + · · · + Xn. Now {N(t)
+1=
n} = {Sn-1::::: t}
n {Sn
> t}
E
:Fn.
3. (Y+, :F) is a submartingale, and T = min{k: Yk :=:: x} is a stopping time. Now 0::::: T 1\ n::::: n, so that JE(Yd) ::::: lE(YfAn) ::::: lE(Y,i), whence
4.
We may suppose that JE(Yo) < oo. With the notation of the previous solution, we have that
5. It suffices to prove that lEYs ::::: lEYT, since the other inequalities are of the same form but with different choices of pairs of stopping times. Letlm be the indicator function of the event {S < m ::::: T}, and define n
Zn =
L
lm(Ym- Ym-1),
m=1
Note that Im is :Fm-1-measurable, so that
399
O:sn :S N.
[12.4.6]-[12.5.2]
Solutions
Martingales
since Y is a submartingale. Therefore E(ZN) ~ E(ZN-1) ZN = Yr- Ys, and therefore E(Yr) ~ E(Ys).
~
· · · ~ E(Zo)
= 0.
On the other hand,
6. De Moivre's martingale is Yn = (q I p )Sn, where q = 1 - p. Now Yn ~ 0, and E(Yo) = 1, and the maximal inequality gives that ll' (max Sm O:=;m:=;n
~ x)
= ll' (max Ym O:=;m:=;n
~ (qlp)x)::: (plq)x.
Take the limit as n -+ oo to find that S00 = supm Sm satisfies 00
L ll'(S
E(Soo) =
~
00
x=l
x)::: _P__ q- P
We can calculate E(S00 ) exactly as follows. It is the case that S00 ~ x if and only if the walk ever visits the point x, an event with probability fx for x ~ 0, where f = pI q (see Exercise (5.3.1)). The inequality of (*) may be replaced by equality. 7.
(a) First,
0
n {T::: n} =
0 E :Fn.
Secondly, if An {T::: n}
E :Fn
then
Ac n {T::: n} = {T::: n} \(An {T::: n}) E :Fn.
Thirdly, if A1, Az, ... satisfy Ai
n {T ::: n} E
:Fn for each i, then
(U Ai) n {T ::: n} = l) (Ai n {T ::: n}) E :Fn. l
l
Therefore :Fr is a a-field. For each integer m, it is the case that {T < n}
{T
E :Fr
ifm > n, ifm::: n,
for all m.
n
(An{S:::T})n{T:Sn}=
U (An{S:::m})n{T=m}, m=O
the union of events in :Fn, which therefore lies in :Fn. Hence A n {S ::: T} E :Fr. (c) We have {S::: T} = n, and (b) implies that A E :Fr whenever A E :Fs.
12.5 Solutions. Optional stopping Under the conditions of (a) or (b), the family {YT/\n : n ~ 0} is uniformly integrable. Now T 1\ n -+ T as n -+ oo, so that Yr An -+ Yr a.s. Using uniform integrability, E(Yr An) -+ E(Yr ), and the claim follows by the fact that E(Yr An) = E(Yo).
1.
2. It suffices to prove that {Yr An : n uniformly integrable if
~
0} is uniformly integrable. Recall that {Xn : n
400
~
0} is
Solutions
Optional stopping
[12.5.3]-[12.5.5]
(a) Now, lE (IYTAnll(IYT/\nl:::aJ) = lE (IYrll(T:sn,IYTI2:aJ) + lE (IYnll(T>n,IYnl:::aJ) ::0 lE (IYrll(lhl:::aJ) + lE (IYnll{T>nJ) = g(a) + h(n),
say. We have that g(a) ----+ 0 as a ----+ oo, since lEIYrl < oo. Also h(n) ----+ 0 as n ----+ oo, so that supn>N h(n) may be made arbitrarily small by suitable choice of N. On the other hand, lE (IYnll(IYnl~aJ)----+ 0 as a----+ oo uniformly inn E {0, 1, ... , N}, and the claim follows. (b) Since Y;i defines a submartingale, we have that supn lE(YfAn) ::0 supn JE(Y;i) < oo, the second inequality following by the uniform integrability of {Yn }. Using the martingale convergence theorem, YT/\n----+ Yr a.s. wherelEIYrl < oo. Now
Also lP'(T > n) ----+ 0 as n ----+ oo, so that the final two terms tend to 0 (by the uniform integrability of the Yi and the finiteness of lEI Yr I respectively). Therefore Yr An standard theorem (7.10.3).
~
Yr, and the claim follows by the
3. By uniform integrability, Y00 = limn-+oo Yn exists a.s. and in mean, and Yn = lE(Y00 I :Fn). (a) On the event {T = n} it is the case that Yr = Yn and JE(Y00 I :Fr) = lE(Y00 I :Fn); for the latter statement, use the definition of conditional expectation. It follows that Yr = JE(Y00 I :Fr ), irrespective of the value of T. (b) We have from Exercise (12.4.7) that :Fs ~ :Fr. Now Ys = JE(Yoo I :Fs) = lE{lE(Y00 I :Fr) I :Fs)
= JE(Yr I :Fs).
4.
Let T be the time until absorption, and note that {Sn} is a bounded, and therefore uniformly integrable, martingale. Also IP'(T < oo) = 1 since T is no larger than the waiting time for N consecutive steps in the same direction. It follows that JE(So) = lE(Sr) = NIP'(Sr = N), so that IP'(Sr = N) = JE(So)/ N. Secondly, {S~- n : n 2':: 0} is a martingale (see Exercise (12.1.4)), and the optional stopping theorem (if it may be applied) gives that lE(S6)
= JE(Sf -
T)
= N 21P'(Sr = N) -lE(T),
and hence JE(T) = NlE(So) - lECS6) as required. It remains to check the conditions of the optional stopping theorem. Certainly IP'(T < oo) = 1, and in addition lE(T 2) < oo by the argument above. We have that lEI Sf- Tl ::0 N 2 + JE(T) < oo. Finally, lE{ (S~ - n)l(T>n}} ::0 (N 2 + n)IP'(T > n) ----+ 0 as n ----+ oo, since JE(T 2) < oo.
5. Let :Fn = cr(SJ, S2, ... , Sn). It is immediate from the identity cos(A + ).) + cos(A- ).) = 2 cos A cos). that lE(Yn+l I :Fn)
=
cos[,J,.(Sn + 1- ~(b- a))]+ cos[,J,.(Sn- 1- ~(b- a))]
2(cos ,J,.)n+l
= Yn,
and therefore Y is a martingale (it is easy to see that lEI Yn I < oo for all n ). SupposethatO <). < rrj(a +b), andnotethatO ::0 I,J,.{Sn- ~(b -a)}l <~).(a +b)< ~JT for n ::0 T. Now Yr An constitutes a martingale which satisfies
cos{~,J,.(a+b)} _ _-=c__=-_ < Yr An < (cos ,J,.)T 1\n
-
401
-
1 · (cos ,J,.)T
[12.5.6]-[12.5.8]
Solutions
Martingales
If we can prove that E{(cos A.)-T} < oo, it will follow that {Yr 1\n} is uniformly integrable. This will imply in tum that E(Yr) = Iimn-+oo E(Yr 1\n) = E(Yo), and therefore cos{~A.(a + b)}E{(cosA.)-T} = cos{iA.(b- a)}
as required. We have from (*) that
Now T
1\
n --+ T as n --+ oo, implying by Fatou's lemma that
E{(cosA.)-T}:::
cos{iA.(a- b)}
E(Yo)
cos{~A.(a +b)}
cos{iA.(a +b)}·
6. (a) The occurrence of the event {U = n} depends on Sr, S2, ... , Sn only, and therefore U is a stopping time. Think of U as the time until the first sequence of five consecutive heads in a sequence of coins tosses. Using the renewal-theory argument of Problem (10.5.17), we find that E(U) = 62. (b) Knowledge of Sr, S2, ... , Sn is insufficient to determine whether or not V = n, and therefore V is not a stopping time. Now E(V) = E(U)- 5 =57. (c) W is a stopping time, since it is a first-passage time. Also E(W) = oo since the walk is null persistent. 7.
With the usual notation,
E(Mm+n I :Fm)
=E(
f
r=O
Sr
+
~
Sr -
~(Sm+n -
Sm
+ Sm) 3
1
:Fm)
r=m+l
+ nSm- SmE{(Sm+n- Sm) 2} = Mm + nSm- nSmE(XI) = Mm.
= Mm
Thus {Mn : n :::: 0} is a martingale, and evidently Tis a stopping time. The conditions of the optional stopping theorem (12.5.1) hold, and therefore, by a result of Example (3.9.6),
8. We partition the sequence into consecutive batches of a + b flips. If any such batch contains only l's, then the game is over. Hence lP'(T > n(a +b))::; {1- (~)a+b}n--+ 0 as n--+ oo. Therefore, EISf- Tl ::; E(Sf) + E(T) ::; (a+ b) 2 + E(T) < oo, and as n--+ oo.
402
Solutions
Problems
[12.7.1]-[12.9.2]
12.7 Solutions. Backward martingales and continuous-time martingales 1.
Lets~ t. WehavethatE(I](X(t))
I J=S, Xs =
i) = '£j Pij(t- s)IJ(j). Hence
I
i)
:tlE(IJ(X(t)) :Fs, Xs =
= (Pr-sGJ7'); = 0,
so that E(I](X (t)) I :Fs, Xs = i) = l](i ), which is to say that lE(IJ (X (t))
2. Let W(t) = exp{-BN(t) + A.t(l- e-li)} where constitutes a martingale. Furthermore
e 0::: 0.
I :Fs) = I](X (s )).
It may be seen that W(t
1\
Ta), t 0::: 0,
where, by assumption, the limit has finite expectation for sufficiently small positive e (this fact may be checked easily). In this case, {W(t 1\ Ta): t 0::: 0} is uniformly integrable. Now W(t 1\ Ta)-+ W(Ta) a.s. as t -+ oo, and it follows by the optional stopping theorem that
Writes = e-li to obtain s-a = E{eHae 1-sl}. Differentiate at s = 1 to find that a = AE(Ta) and a(a + 1) =A 2E(T}), whence the claim is immediate. 3. Let 9-m be the a -field generated by the two sequences of random variables Sm, Sm+ 1 ... , Sn and Um+1, Um+2• ... , Un. It is a straightforward exercise in conditional density functions to see that -1
E (Um+1
!oUm+2 I 9-m+1) -- 0
(m
+ l)xm- 1 d x -m-+- 1
(U
m+2
)m+1
-
U ' m m+2
whence E(Rm I 9-m+ 1) = Rm+1 as required. [The integrability condition is elementary.] LetT= max{m : Rm 0::: 1} with the convention that T = l if Rm < 1 for all m. As in the closely related Example (12.7.6), Tis a stopping time. We apply the optional stopping theorem (12.7.5) to the backward martingale R to obtain that E(RT I 9-n) = Rn = Sn/t. Now, RT 0::: 1 on the event {Rm 0::: l for some m ~ n}, whence
~ = E(RT I Sn = t
y) 0::: IP'(Rm 0::: l forsomem
~ n I Sn = y).
[Equality may be shown to hold. See Karlin and Taylor 1981, pages 110-113, and Example (12.7.6).]
12.9 Solutions to problems 1.
Clearly E(Zn)
~
(M + m )n, and hence El Yn I < oo. Secondly, Zn+ 1 may be expressed as
'£f,:; 1 X;+ A, where X1, X2, ... are the family sizes of the members of the nth generation, and A is the number of immigrants to the (n + l)th generation. Therefore E(Zn+1
I Zn) = tJ-Zn +
m, whence
1 { MZn + m ( 1- 1 -l _Mn+1)} E(Yn+1 I Zn) = Mn+ M = Yn. 1 2. Each birth in the (n + 1)th generation is to an individual, say the sth, in the nth generation. Hence, for each r, Ben+ 1),r may be expressed in the form Ben+ 1),r = Bn,s + Bj(s ), where Bj (s) is the age of the parent when its jth child is born. Therefore E{
L e-liB(n+l).r I:Fn} = E{ ~ e-lieBn.s+Bjes)) I:Fn} = L e-liBn.s M1 (B), r
s
S,J
403
[12.9.3]-[12.9.5]
Solutions
which gives that E(Yn+ 1 I :Fn) 3.
Martingales
=
Yn. Finally, E(Y1 (e))
=
1, and hence E(Yn (e))
=
1.
If x, c > 0, then lP' (max Yk > 1~k~n
x) ::=: lP' (max (Yk + c) > (x + c) 2
1~k~n
2 ).
Now (Yk + c) 2 is a convex function of Yk, and therefore defines a submartingale (Exercise (12.1.7)). Applying the maximalinequality to this submartingale, we obtain an upper bound ofE{ (Yn +c) 2 }/ (x + c) 2 for the right-hand side of(*). We set c = E(Y1)/x to obtain the result.
4. (a) Note that Zn = Zn-1 + Cn{Xn- E(Xn I :Fn_I)}, so that (Z, !F) is a martingale. LetT be the stopping timeT= min{k: qYk 2:: x}. ThenE(ZTAn) = E(Zo) = 0, so that
since the final term in the definition of Zn is non-negative. Therefore n
xlP'(T ::S n) ::S E{cTAnyT/\n} ::S LckE{E(Xk I :Fk-d},
k=1 where we have used the facts that Yn 2:: 0 and E(Xk I :Fk-d 2:: 0. The claim follows. (b) Let X 1, X 2, ... be independent random variables, with zero means and finite variances, and let Yj = 1 X;. Then Y/ defines a non-negative submartingale, whence
L;{=
5. The function h(u) = lulr is convex, and therefore Y;(m) = IS;- Smlr, i 2:: m, defines a submartingale with respect to the filtration :F; = a ({x1 : 1 :::: j :::: i}). Apply the HRC inequality of Problem (12.9.4), with q = 1, to obtain the required inequality. If r = 1, we have that
E(ISm+n- Sml) ::S
m+n L EIZkl k=m+1
by the triangle inequality. Let m, n --+ oo to find, in the usual way, that the sequence {Sn} converges a.s.; Kronecker's lemma (see Exercise (7.8.2)) then yields the final claim. Suppose 1 < r ::=: 2, in which case a little more work is required. The function h is differentiable, and therefore
h(v)- h(u) = (v- u)h 1 (u)
+ fov-u {h'(u +x)- h 1(u)} dx.
Now h'(y) = rlylr- 1sign(y) has a derivative decreasing in IYI· It follows (draw a picture) that h 1 (u + x)- h 1(u) ::=: 2h 1 ( ix) if x 2:: 0, and therefore the above integral is no larger than 2h(i(v- u)). Apply this with v = Sm+k+1 - Sm and u = Sm+k- Sm, to obtain
404
Solutions [12.9.6]-[12.9.10]
Problems
Sum over k and use the fact that
to deduce that E(ISm+n- smn
~ 2 2 -r
m+n
L
E(lzkn·
k=m+l
The argument is completed as after (*). 6.
With
h
= l{Yk=O)• we have that
+ nYn-tiXn 1(1 - ln-t) I.Tn-1) = ln-tE(Xn) + nYn-1 (1- ln-t)EIXnl = Yn-t
E(Yn I .Tn-t) = E( Xnln-1
since E(Xn) = 0, EIXnl = n- 1. Also EIYnl ~ E{IXniO El Yn I < oo. Therefore (Y, !F) is a martingale. Now Yn
=
0 if and only if Xn
=
0. Therefore lP'(Yn
+ niYn-tl)} =
0)
=
and EIYtl < oo, whence
lP'(Xn
=
0)
=
1 - n- 1 ---+ 1
as n ---+ oo, implying that Yn ~ 0. On the other hand, Ln lP'(Xn =1- 0) = oo, and therefore lP'(Yn =1- 0 i.o.) = 1 by the second Borel-Cantelli lemma. However, Yn takes only integer values, and therefore Yn does not converge to 0 a.s. The martingale convergence theorem is inapplicable since supn EIYnl = 00. 7. Assume that t > 0 and M(t) = 1. Then Yn = e 1Sn defines a positive martingale (with mean 1) with respect to .Tn = a(Xt, X2, ... , Xn). By the maximal inequality,
and the result follows by taking the limit as n ---+ oo.
8. The sequence Yn = ~Zn defines a martingale; this may be seen easily, as in the solution to Exercise (12.1.15). Now {Yn} is uniformly bounded, and therefore Y 00 = limn---+oo Yn exists a.s. and satisfies E(Y00 ) = E(Yo) = ~. Suppose 0 < ~ < 1. In this case Zt is not a.s. zero, so that Zn cannot converge a.s. to a constant c unless c E {0, oo}. Therefore the a.s. convergence of Yn entails the a.s. convergence of Zn to a limit random variable taking values 0 and oo. In this case, E(Y00 ) = 1 · lP'(Zn ---+ 0) + 0 · lP'(Zn ---+ oo ), implying that lP'(Zn ---+ 0) = ~, and therefore lP'(Zn ---+ oo) = 1 - ~.
9. It is a consequence of the maximal inequality that lP'(Y; =:: x) ~ x- 1E(Ynl{Y,i::::x)) for x > 0. Therefore E(Y;) = fooo lP'(Y; =:: x) dx
~ 1 + E { Yn
1
00
~ 1+
1
00
lP'(Y;
=:: x) dx
x-l I(l,Y,t](x) dx}
= 1 + E(Yn log+ Y;) ~ 1 + E(Yn log+ Yn)
+ E(Y;)fe.
10. (a) We have, as in Exercise (12.7.1), that E(h(X(t)) I B,X(s)
=i) =
LPij(t)h(j) j
405
fors < t,
[12.9.11]-[12.9.13]
Solutions
Martingales
for any event B defined in terms of {X (u) : u ::=: s}. The derivative of this expression, with respect to t, is (P1 Gh')i, where P1 is the transition semigroup, G is the generator, and h = (h(j) : j :::: 0). In this case,
(Gh1)j
=L
gjkh(k)
= Aj { h(j + 1)- h(j)}- IJ-j { h(j)- h(j- 1)} = 0
k
for all j. Therefore the left side of (*) is constant for t ::=: s, and is equal to its value at time s, i.e. X (s). Hence h(X(t)) defines a martingale. (b) We apply the optional stopping theorem with T = min{t: X(t) E {0, n}} toobtainE(h(X(T))) = E(h(X(O))), and therefore (1 - n(m))h(n) = h(m) as required. It is necessary but not difficult to check the conditions of the optional stopping theorem.
11. (a) Since Y is a submartingale, so is y+ (see Exercise (12.1.6)). Now
Therefore {E(Y:+m I :Fn) : m :::: 0} is (a.s.) non-decreasing, and therefore converges (a.s.) to a limit Mn. Also, by monotone convergence of conditional expectation,
E(Mn+l
I :Fn) = m--+oo lim E{E(Y:+m+l I :Fn+I) I:Fn} = m--+00 lim E(Y:+m+l I :Fn) = Mn,
and furthermore E(Mn) = limm---+oo E(Y:+n) :::0 M. It is the case that Mn is :Fn-measurable, and therefore it is a martingale. (b) We have that Zn = Mn - Yn is the difference of a martingale and a submartingale, and is therefore a supermauingale. Also Mn ::=: Y: ::=: 0, and the decomposition for Yn follows. (c) In thi~ case Zn is a martingale, being the difference of two martingales. Also Mn :::: E(Y: I :Fn) = Y: ::=: Yn a.s., and the claim follows.
12. We may as well assume that JJ- < P since the inequality is trivial otherwise. The moment generating function of P- Cr is M(t) = iCP-JLH5;a 212 , and we choose t such that M(t) = l, i.e., t = -2(P- J.J-)/a 2. Now define Zn = min{etYn, 1} and :Fn = a(Cr, C2, ... , Cn). Certainly EIZnl < oo; also E(Zn+l I :Fn) :::0 E(etYn+l I :Fn) = etYn M(t) = etYn and E(Zn+l I :Fn) :S 1, implying that E(Zn+l I :Fn) :S Zn. Therefore (Zn, :Fn) is a positive supermartingale. LetT = inf{n : Yn ::=: 0} = inf{n : Zn = 1}. Then T 1\ m is a bounded stopping time, whence E(Zo) ::=: E(ZT Am) ::=: JP>(T ::=: m). Let m ---+ oo to obtain the result.
13. Let :Fn = a(Rr, R2, ... , Rn). (a) 0 :S Yn :S 1, and Yn is :Fn-measurable. Also
E(Rn+l
I Rn)
= Rn
Rn
+ n+r+ b'
whence Yn satisfies E(Yn+I I :Fn) = Yn. Therefore {Yn : n ::=: 0} is a uniformly integrable martingale, and therefore converges a.s. and in mean. (b) In order to apply the optional stopping theorem, it suffices that JP>(T < oo) = l (since Y is uniformly integrable). However JP>(T > n) = ~ n~I = (n + l)- 1 ---+ 0. Using that theorem,
1· ··· E(YT) = E(Yo), which is to say that E{T I (T + 2)} = 1, and the result follows. (c) Apply the maximal inequality.
406
Solutions [12.9.14]-[12.9.17]
Problems
14. As in the previous solution, with fi,n the a-field generated by A1, A2, ... and :Fn, E(Yn+1
I 13 (J>n) =
Rn + An ) ( Rn ) Rn + Bn +An Rn + Bn Rn ---=Yn, Rn +Bn (
+(
Rn
Rn
+ Bn + An
) (
Bn Rn
+ Bn
)
sothatE(Yn+1 I :Fn) = E{E(Yn+1 I fi,n) I :Fn} = Yn. Also IYn I ::=: 1, and therefore Yn is a martingale. We need to show that lP'(T < oo) = 1. Letln be the indicator function of the event {T > n}. We have by conditioning on the An that
E(In
I A)
IT
= n-1 ( 1 - -1- ) --+
J=O
2 + SJ
IT 00
J=O
(
1 - -1- ) 2 + S1
L-{=
as n --+ oo, where s1 = 1 Ai. The infinite product equals 0 a.s. if and only if"L,1 (2+ SJ )- 1 = oo a.s. By monotone convergence, lP'(T < oo) = 1 under this condition. If this holds, we may apply the optional stopping theorem to obtain that E(YT) = E(Yo), which is to say that
15. At each stage k, let Lk be the length of the sequence 'in play', and let Yk be the sum of its entries, so that Lo = n, Yo = L,j= 1 Xi. If you lose the (k + 1)th gamble, then Lk+ 1 = Lk + 1 and Yk+1 = Yk + Zk where Zk is the stake on that play, whereas if you win, then Lk+1 = Lk - 2 and Yk+1 = Yk- Zk; we have assumed that Lk :::: 2, similar relations being valid if Lk = 1. Note that Lk is a random walk with mean step size -1, implying that the first-passage time T to 0 is a.s. finite, and has all moments finite. Your profits at time k amount to Yo - Yb whence your profit at time T is Yo, since YT = 0. Since the games are fair, Yk constitutes a martingale. Therefore E(YT 1\m) = E(Yo) =j:. 0 for all m. However T 1\ m --+ T a.s. as m --+ oo, so that YT 1\m --+ YT a.s. Now E(YT) = 0 =j:. Iimm---+ooE(YT/\m), and it follows that {YT/\m : m :0:: 1} is not uniformly integrable. Therefore E(supm YT 1\m) = oo; see Exercise (7.10.6). 16. Since the game is fair, E(Sn+1 I Sn) = Sn. Also ISnl ::=: 1 + 2 + · · · + n < oo. Therefore Sn is a martingale. The occurrence of the event {N = n} depends only on the outcomes of the coin-tosses up to and including the nth; therefore N is a stopping time. A tail appeared at time N- 3, followed by three heads. Therefore the gamblers G1, G2, ... , G N -3 have forfeited their initial capital by time N, while G N -i has had i + 1 successful rounds for 0 ::=: i ::=: 2. Therefore SN = N- (p- 1 + p- 2 + p- 3 ), after a little calculation. It is easy to check that N satisfies the conditions of the optional stopping theorem, and it follows that E(SN) = E(So) = 0, which is to saythatE(N) = p-1 + p-2 + p-3.
In order to deal with HTH, the gamblers are re-programmed to act as follows. If they win on their first bet, they bet their current fortune on tails, returning to heads thereafter. In this case, SN = N- (p- 1 + p- 2q- 1) where q = 1- p (remember that the game is fair), and therefore E(N) = p-1 + p-2q-1.
17. Let :Fn = a ({Xi , Yi : 1 ::=: i ::=: n}), and note that T is a stopping time with respect to this filtration. Furthermore lP'(T < oo) = 1 since Tis no larger than the first-passage time to 0 of either of the two single-coordinate random walks, each of which has mean 0 and is therefore persistent. Let a'f = var(X1) and ai = var(YI). We have that Un - Uo and Vn - Vo are sums of independent summands with means 0 and variances a'f and ai respectively. It follows by considering
407
[12.9.18]-[12.9.19]
Solutions
Martingales
the martingales (Un- Uo) 2 -no} and (Vn- Vo) 2 -no} (see equation (12.5.14) and Exercise (10.2.2)) that E{(UT - Uo) 2} = o}E(T), E{(VT - Vo) 2} = o}E(T). Applying the same argument to (Un E { (UT
+ VT- Uo-
+ Vn)- (Uo + Vo), we obtain
Vo) 2} = E(T)E{(Xr
+ Yr) 2} =
E(T)(a12 + 2c + a 22).
Subtract the two earlier equations to obtain E{ (UT - Uo)(VT - Vo)} = cE(T) if E(T) < oo. Now UT VT = 0, and in addition E(UT) = Uo, E(VT) = Vo, by Wald's equation and the fact that E(X r) = E(Yr) = 0. It follows that -E(Uo Vo) = cE(T) if E(T) < oo, in which case c < 0. Suppose conversely that c < 0. Then(*) is valid with T replaced throughout by the bounded stopping timeT 1\ m, and hence
0 :S E(UT Am VTAm) = E(Uo Vo)
+ cE(T 1\ m).
Therefore E(T 1\ m) :S E(Uo Vo)/(21cl) for all m, implying that E(T) = limm-+oo E(T and so E(T) = -E(Uo Vo)/c as before.
1\
m) < oo,
18. Certainly 0 :S Xn :S 1, and in addition Xn is measurable with respect to the a-field :Fn = a(Rr, R2, ... , Rn). Also E(Rn+l I Rn) = Rn- Rn/(52- n), whence E(Xn+l I :Fn) = Xn. Therefore Xn is a martingale. A strategy corresponds to a stopping time. If the player decides to call at the stopping time T, he wins with (conditional) probability X T· and therefore IP'(wins) = E(XT ), which equals E(Xo) (= by the optional stopping theorem. Here is a trivial solution to the problem. It may be seen that the chance of winning is the same for a player who, after calling "Red Now", picks the card placed at the bottom of the pack rather than that at the top. The bottom card is red with probability irrespective of the strategy of the player.
i)
i,
19. (a) A sums of money in weekt is equivalentto a sums /(1 +a)t in weekO, since the latter sum may be invested now to yield s in week t. If he sells in week t, his discounted costs are :L~= 1 cI (1 + a )n and his discounted profit is Xt/(1 + a)t. He wishes to find a stopping time for which his mean discounted gain is a maximum. Now
so that tt(T) = E{ (1
+ a)-T ZT} -
(c/a).
(b) The function h(y) = ay - fy IP'(Zn > y) dy is continuous and strictly increasing on [0, 100 ), with h(O) = -E(Zn) < 0 and h(y) --+ oo as y --+ oo. Therefore there exists a unique y (> 0) such that h (y) = 0, and we choose y accordingly. (c) Let :Fn = a(Zr, Z2, ... , Zn). We have that 00
E(max{Zn, y}) = y
+
l
00
[1- G(y)]dy = (1 +a)y
where G(y) = IP'(Zn :S y). Therefore E(Vn+l I :Fn) = (1 non-negative supermartingale.
408
+ a)-ny
:S Vn, so that CVn, :Fn) is a
Solutions [12.9.20]-[12.9.21]
Problems
Let ~-t(r) be the mean gain of following the strategy 'accept the first offer exceeding r - (cja)'. The corresponding stopping timeT satisfies JP>(T = n) = G(r)n(l- G(r)), and therefore 00
~-t(r)
+ (cja)
=
L E{ (1 + a)-T Zr l{T=n)} n=O 00
= L(l
+ a)-nG(r)n(l- G(r))E(Z, I z,
n=O
=
1 +a { r(l - G(r)) 1 +a- G(r)
+
1
00
,
> r)
(1 - G(y)) dy } .
Differentiate with care to find that the only value of r lying in the support of Z 1 such that ~-t' (r) = 0 is the value r = y. Furthermore this value gives a maximum for ~-t( r ). Therefore, amongst strategies of the above sort, the best is that with r = y. Note that ~-t(Y) = y(l +a) - (cja). Consider now a general strategy with corresponding stopping time T, where JP>(T < oo) = 1. For anypositiveintegerm, T /\misaboundedstoppingtime, whenceE(VTAm):::: E(Vo) = y(l+a). Now IVT Am I :::: l::~o IVil, and l::~o EIVi I < oo. Therefore {Vr Am : m :=::: 0} is uniformly integrable. Also Vr Am --+ Vr a.s. as m --+ oo, and it follows that E(Vr Am) --+ E(Vr ). We conclude that J-t(T) = E(Vr)- (cja) :::: y(l +a)- (cja) = J-t(y). Therefore the strategy given above is optimal. (d) In the special case, JP>(ZI > y) = (y - 1)- 2 for y ::::, 2, whence y = 10. The target price is therefore 9, and the mean number of weeks before selling is G(y)/(1 - G(y)) = 80.
20. Since G is convex on [0, oo) wherever it is finite, and since G(l) =land G'(l) < 1, there exists a unique value of IJ (> 1) such that G(l]) = IJ. Furthermore, Yn = IJZn defines a martingale with mean E(Yo) = IJ. Using the maximal inequality (12.6.6),
n
1
k
IJ) ::S
JP>(supZn::::, k) = lP' ( supYn :=::: n
k-l I]
for positive integers k. Therefore 00
E(supZn) ::S n
1
L - .1
k=I I] -
21. LetMn bethenumberpresentafterthenthround, soMo = K, andMn+I = Mn -Xn+I· n::::, 1, where Xn is the number of matches in the nth round. By the result of Problem (3.11.17), EXn = 1 for all n, whence E(Mn+I + n + 1 I :Fn) = Mn + n, where :Fn is the a-field generated by Mo, M1, ... , Mn. Thus the sequence {Mn + n} is a martingale. Now, N is clearly a stopping time, and therefore K = Mo + 0 = E(MN + N) =EN. We have that
+ n + 1) 2 + Mn+I I Fn} = (Mn + n) 2 - 2(Mn + n)E(Xn+I ::S (Mn + n) 2 + Mn,
E{ (Mn+I
- 1) + Mn
+ E{ (Xn+I
where we have used the fact that
1 if Mn > 1, if Mn = 1.
var(Xn+l I :Fn) = { 0
409
I
- 1) 2 - Xn+I :Fn}
[12.9.22]-[12.9.24]
Solutions
Martingales
+ n) 2 + Mn}
Hence the sequence {(Mn supermartingales,
is a supermartingale. By an optional stopping theorem for
and therefore var(N) _:::: K.
22. In the usual notation, E(M(s
+ t) I.'Fs)
= E(fos W(u)du
+ 1s+t W(u)du-
HW(s
= M(s) + tW(s)- W(s)E([W(s + t)-
+ t)- W(s) + W(s)} 3 1.'Fs)
W(s)l 2
I.'Fs) = M(s)
as required. We apply the optional stopping theorem (12.7.12) with the stopping timeT= inf{ u : W ( u) E {a, b}}. The hypotheses of the theorem follow easily from the boundedness of the process fort E [0, T], and it follows that
· Hence the required area A has mean E(A) = E( (T W(u)du) =
h
3 ) = ~a 3 (~) + ~b 3 (-a-). ~E(W(T) 3 3 a-b 3 a-b
[We have used the optional stopping theorem twice actually, in that E(W(T)) = 0 and therefore JP>(W(T) =a)= -bj(a- b).]
23. With .'Fs = a(W(u) : 0 _:::: u _:::: s), we have for s < t that E(R(t) 2 I .'Fs) = E(IW(s)l 2 + IW(t)- W(s)l 2
+ 2W(s) · (W(t)-
W(s)) I:Fs) = R(s) 2
+ (t- s),
and the first claim follows. We apply the optional stopping theorem (12.7.12) with T = inf{u : IW(u)l =a}, as in Problem (12.9.22), to find that 0 = E(R(T) 2 - T) = a 2 - E(T).
24. We apply the optional stopping theorem to the martingale W(t) with the stopping timeT to find that E(W(T)) = -a(1 - Pb) + bpb = 0, where Pb = JP>(W(T) = b). By Example (12.7.10), W (t) 2 - t is a martingale, and therefore, by the optional stopping theorem again,
whence E(T) = ab. For the final part, we take a = band apply the optional stopping theorem to the 2 t] to obtain martingale exp[eW(t)-
ie
on noting that the conditional distribution ofT given W (T) = b is the same as that given W (T) = -b. Therefore, E(e-i 02 T)
=
1/ cosh(be), and the answer follows by substituting s
410
=
ie 2.
13 Diffusion processes
13.3 Solutions. Diffusion processes 1.
It is easily seen that
I
E{ X(t +h)- X(t) X(t)} =(A.- p,)X(t)h + o(h), E( {X (t +h) - X (t)} 2 1 X (t)) = (A.+ p,)X (t)h + o(h),
which suggest a diffusion approximation with instantaneous mean a(t, x) = (A.- p,)x and instantaneous variance b(t, x) =(A.+ p,)x. 2. The following method is not entirely rigorous (it is an argument of the following well-known type: it is valid when it works, and not otherwise). We have that
-aM = at
1oo e -oo
0Yaf -
at
dy
=
1oo {ea(t, y) + ~e b(t, y) }e -oo 2
0Y f
dy,
by using the forward equation and integrating by parts. Assume that a(t, y) = I:n an (t)yn, b(t, y) = I:n f3n (t)yn. The required expression follows from the 'fact' that
3.
Using Exercise (13.3.2) or otherwise, we obtain the equation aM= emM + le 2M 2 at
with boundary condition M(O, e) = l. The solution is M(t) = exp{ ~e (2m + e)t}.
4.
Using Exercise (13.3.2) or otherwise, we obtain the equation aM aM 1 2 -=-e-+-e M 2 ae at
with boundary condition M(O, e) = l. The characteristics of the equation are given by dt 1
de e
2dM e2M'
with solution M(t, e) = e! 02 g(ee- 1) where g is a function satisfying 1 = e! 02 g(e). Therefore M = exp{~e 2 (1- e- 21 )}.
411
[13.3.5]-[13.3.10]
Solutions
Diffusion processes
5. Fix t > 0. Suppose we are given W1 (s), W2(s), W3(s), for 0:::;: s :::;: t. By Pythagoras's theorem, R(t + u) 2 =Xi+ X~+ X~ where the Xi are independent N(Wi (t), u) variables. Using the result of Exercise (5.7.7), the conditional distribution of R(t + u) 2 (and hence of R(t + u) also) depends only on the value of the non-centrality parameter e = R(t) 2 of the relevant non-central x2 distribution. It follows that R satisfies the Markov property. This argument is valid for the n-dimensional Bessel process. 6. By the spherical symmetry of the process, the conditional distribution of R(s +a) given R(s) = x is the same as that given W(s) = (x, 0, 0). Therefore, recalling the solution to Exercise (13.3.5),
lP'(R(s +a):::;: y I R(s) =x) =
=
I
i { (u-x) 2 +v 2 +w 2 } (u,v,w): dudvdw 2a 2:rra 312 exp u2+v2+w2:sy2 ( )
y 1 2:n: l:n: 1 { 1p=O 312 exp ¢=0 0=0 (2:rra)
=loy~ { exp (
(p
~ax)2) -
p 2 - 2px cos e + x 2 } p 2 sm8d8d¢dp . 2a exp (
(p ;ax)2)} dp,
and the result follows by differentiating with respect to y. 7. Continuous functions of continuous functions are continuous. The Markov property is preserved because gO is single-valued with a unique inverse. I 2
8. (a) Since E(eaW(t)) = e2a t, this is not a martingale. (b) This is a Wiener process (see Problem (13.12.1)), and is certainly a martingale. (c) With :Ft = a(W(s) : 0:::;: s :::;: t) and t, u > 0, E{ (t + u)W(t + u)- lot+u W(s)ds
I:Ft} =
(t + u)W(t)- lot W(s)ds -lt+u W(t)ds
= tW(t)- lot W(s)ds,
whence this is a martingale. [The integrability condition is easily verified.]
9. (a) With s < t, S(t) = S(s) exp{a(t- s) + b(W(t)- W(s)) }. Now W(t)- W(s) is independent of {W(u) : 0:::;: u :::;: s}, and the claim follows. (b) S (t) is clearly integrable and adapted to the filtration !F = (:Ft) so that, for s < t, E(S(t) I Fs) = S(s)E(exp{a(t- s) + b(W(t)- W(s))} I Fs) = S(s)exp{a(t- s) + ~b 2 (t- s)}, which equals S(s) if and only if a+ ~b 2 = 0. In this case, E(S(t)) = E(S(O)) = 1.
10. Either find the instantaneous mean and variance, and solve the forward equation, or argue directly as follows. With s < t, lP'(S(t):::;: y I S(s) = x) = lP'(bW(t):::;: -at+ logy I bW(s) =-as+ logx). Now b(W(t)- W(s)) is independent of W(s) and is distributed as N(O, b2(t- s)), and we obtain on differentiating with respect to y that
f (t, y 1 s, x) -_
1 ( (log(y/x)-a(t-s)) 2 ) exp , 2 yy2:rrb2(t- s) 2b (t- s)
412
x,y > 0.
Solutions
Excursions and the Brownian bridge
[13.4.1]-[13.6.1]
13.4 Solutions. First passage times 1.
Certainly X has continuous sample paths, and in addition EIX (t) I < oo. Also, if s < t,
as required, where we have used the fact that W(t)- W(s) is N(O, t - s) and is independent of :Fs.
2.
Apply the optional stopping theorem to the martingale X of Exercise (13.4.1), with the stopping time T, to obtain E(X(T)) = 1. Now W(T) = aT + b, and therefore E(elfrT+ilib) = l where 1/J = i ae + 2 . Solve to find that
1e
E(elfrT)
= e-ilib = exp { -b( Jaz- 21/1 +a)}
is the solution which gives a moment generating function. We have that T :::: u if and only if there is no zero in (u, t], an event with probability l (2/n) cos- 1 { .JU?t}, and the claim follows on drawing a triangle.
3.
13.5 Solution. Barriers 1. Solving the forward equation subject to the appropriate boundary conditions, we obtain as usual that F(t, y) = g(t, y I d)+ e-Zmd g(t, y I -d)-
j_:
2me 2 mx g(t, y I x) dx
I
where g(t, y I x) = (2nt) -2 exp{ -(y - x - mt) 2 j(2t)}. The first two terms tend to 0 as t -+ oo, regardless of the sign of m. As for the integral, make the substitution u = (x - y- mt) I ..fi to obtain, as t-+ oo, -
1
-(d+y+mt)/../i
-00
I 2
e-zu { 21mle-21mly 2me 2my - - du -+ ..,fiir 0
ifm < 0, ifm:::: 0.
13.6 Solutions. Excursions and the Brownian bridge 1.
I
2
Let f(t, x) = (2nt)-ze-x /(Zt). It may be seen that
IP'(W(t)
>xI Z, W(O) =
0)
= lim IP'(W(t) w-!-0
>xI Z, W(O) =
w)
where Z = {no zeros in (0, t]}; the small missing step here may be filled by conditioning instead on the event {W(E) = w, no zeros in (E, t]}, and taking the limit as E .j, 0. Now, if w > 0,
IP'(W(t) > x, Z I W(O)
= w) =
1
00
{J(t, y- w)- f(t, y
by the reflection principle, and
IP'(Z I W(O) =
w)
= l- 2
Loo f(t, y)dy 413
=
i:
+ w)} dy
f(t, y)dy
[13.6.2]-[13.6.5]
Solutions
Diffusion processes
by a consideration of the minimum value of Won (0, t]. It follows that the density function of W(t), conditional on Z n {W(O) = w}, where w > 0, is
h ( ) _ f(t, x- w)- f(t, x X -w f(t, y) dy
w
Jw
+ w)
'
X> 0.
Divide top and bottom by 2w, and take the limit as w {. 0: lim hw(x) = _ _1_ 8f = ::_e-x2/(2t) f(t, 0) ax t ,
X> 0.
w-),0
2.
It is a standard exercise that, for a Wiener process W,
C=;),
E{W(t) I W(s) =a, W(1) =
0}
=a
E{W(s) 2 W(O) = W(l) =
0}
= s(l- s),
1
if 0 ::::: s ::::: t ::::: 1. Therefore the Brownian bridge B satisfies, for 0 ::::: s ::::: t ::::: 1, E(B(s)B(t)) = E{ B(s)E(B(t) I B(s))} = 1 - t E(B(s) 2 ) = s(l- t) 1-s as required. Certainly E(B(s)) = 0 for all s, by symmetry. 3. W is a zero-mean Gaussian process on [0, 1] with continuous sample paths, and also W(O) = W(l) = 0. Therefore W is a Brownian bridge if it has the same autocovariance function as the Brownian bridge, that is, c(s, t) = min{s, t}- st. For s < t, cov(W(s), W(t)) = cov(W(s)- sW(1), W(t)- tW(l)) = s- ts- st
+ st =s-st
since cov(W(u), W(v)) = min{u, v}. The claim follows. 4. Either calculate the instantaneous mean and variance of W, or repeat the argument in the solution to Exercise (13.6.3). The only complication in this case is the necessity to show that W(t) is a.s. continuous at t = l, i.e., that u- 1 W(u- l) ---+ 0 a.s. as u---+ oo. There are various ways to show this. Certainly it is true in the limit as u---+ oo through the integers, since, for integral u, W(u- l) may be expressed as the sumofu -1 independent N(O, 1) variables (use the strong law). It remains to fill in the gaps. Let n be a positive integer, let x > 0, and write Mn = max{IW(u)- W(n)l : n::::: u ::::: n + 1}. We have by the stationarity of the increments that oo
oo
E~)
n=O
n=O
X
L)"CMn ::=: nx) = ~)P'CM1 ::=: nx)::::: 1 + - - < oo, implying by the Borel-Cantelli lemma that n- 1 Mn ::::: x for all but finitely many values of n, a.s. Therefore n - 1 Mn ---+ 0 a.s. as n ---+ oo, implying that
. 1 . 1 hm --IW(u)l::::: hm -{IW(n)l +1 n-+oo n
u-+oo u
+ Mn}---+ 0
5.
a.s.
In the notation of Exercise (13.6.4), we are asked to calculate the probability that W has no zeros in the time interval between s /(1 - s) and t /(1 - t). By Theorem (13.4.8), this equals 1 2 1 - - cos-
:rr
s(1-y) 2 -1 t(1-s) =;cos
414
~
V~·
Solutions
Stochastic calculus
[13.7.1]-[13.7.5]
13.7 Solutions. Stochastic calculus 1. Let :Fs = u(Wu : 0:::;: u :::;: s). Fix n :::0: 1 and define Xn(k) = IWktjznl for 0:::;: k:::;: 2n. By Jensen's inequality, the sequence {Xn (k) : 0 :::;: k :::;: 2n} is a non-negative submartingale with respect to the filtration :Fktf2n, with finite variance. Hence, by Exercise (4.3.3) and equation (12.6.2), X~= max{Xn(k): 0 ::=:: k ::=:: 2n} satisfies
E(X~2 ) = 2lo xlP'(X~ > x) dx :::;: 2lo 00
00
E(Wt
= 2E(Wt X~) ::=:: 2VECWhE(Xj?)
/{X~:o:x}) dx =
2E{ wt lox:; dx}
by the Cauchy-Schwarz inequality.
Hence E(X~ 2 ) :::;: 4E(Wh. Now X~ 2 is monotone increasing inn, and W has continuous sample paths. By monotone convergence,
2.
See the solution to Exercise (8.5.4).
3.
(a) We have that
w?,
The first summation equals by successive concellation, and the mean-square limit of the second it in mean square. summation is t, by Exercise (8.5.4). Hence limn--+oo Ir (n) = i Likewise, we obtain the mean-square limits: lim /z(n) = i
n--+oo
4.
w?-
w? +it,
lim /3(n) = lim h(n) =
n--+oo
n--+oo
iw?.
Clearly E(U (t)) = 0. The process U is Gaussian with autocovariance function
Thus U is a stationary Gaussian Markov process, namely the Omstein-Uhlenbeck process. [See Example (9.6.10).] 5.
Clearly E(Ut) = 0. For s < t, E(UsUs+t) = E(Ws Wt)
+ ,B 2 E (1~ 0 1~ 0 e-.B(s-u)Wue-.B(t-v)Wv du dv)
- E ( Wt.B los e-.B(s-u)Wu du) - E ( Ws lot e-.B(t-v)Wv dv) =s+,B 2e-.B(s+t)1s
1 1
e.B(u+v)min{u,v}dudv
u=O v=O
- ,B los e-.B(s-u) min{u, t} du- lot e-,B(t-v) min{s, v} dv e2.Bs - 1 -,B(s+t) 2,8 e
415
[13.8.1]-[13.8.1]
Solutions
Diffusion processes
after prolonged integration. By the linearity of the definition of U, it is a Gaussian process. From the calculation above, it has autocovariance function c(s, s +t) = (e-f3(t-s) -e-f3(t+sl)j(2(3). From this we may calculate the instantaneous mean and variance, and thus we recognize an Omstein-Uhlenbeck process. See also Exercise (13.3.4) and Problem (13.12.4).
13.8 Solutions. The Ito integral 1. (a) Fix t > 0 and let n :=:: I and 8 = tjn. We write tj = jtjn and Vj = Wtr By the absence of correlation of Wiener increments, and the Cauchy-Schwarz inequality,
~ ~{(tj+1- tj) ltj+l E(IVH1- Wsl 2)ds} j=O
=
tl
n-1 1 -(tj+1 -tj) 3 . 0 2
L
=
n-1 1 (t)3
L. 2 n
--+ 0
asn--+ oo .
]= 0
]=
Therefore,
=
lim ( tWtn--'>00
n-1
L vj+1 (tj+1- tj) )
= tWt-
j=O
lor Ws ds. t
(b) As n --+ oo, n-1 n-1 2:: V/(Vi+1- Vj) = 2:: {v/+ 1 - V/- 3Vj(Vj+1 - Vj) 2 -
t
t w?- L [
--+
t
tw?- lot W(s)ds+O+O.
The fact that the last two terms tend to 0 in mean square may be verified in the usual way. For example,
E(I~(Yj+1- Yj) 3 J=O
n ~E[(Vj+l=
J=O n-1 = 6 (tj+ 1 j=O
L
Vj) 6] n-1
fj ) 3 =
416
6
L U·) j=O n
3 --+ 0
as n--+ oo.
Solutions
Ito'sjormula (c) It was shown in Exercise (13.7.3a) that f~ Ws dWs = i
[13.8.2]-[13.9.1]
w?- it. Hence,
and the result follows because IECWh = t. 2. Fix t > 0 and n 2: 1, and let 8 = t/n. We set Vj = Wjtfn· It is the case that Xt = limn---+oo L,j Vj(tj+l- tj). Each term in the sum is normally distributed, and all partial sums are multivariate normal for all8 > 0, and hence also in the limit as 8 --+ 0. Obviously IE(Xt) = 0. For s ::::: t, IE(XsXt) =lot los IE(Wu Wv) du dv =lot los min{u, v} du dv =
losl-u 2 du + los u(t0 2
0
u) du = s 2
(t2
- - -s) .
6
Hence var(X t) = ~ t 3 , and the autocovariance function is p(Xs, Xt) = 3
3.
By the Cauchy-Schwarz inequality, as n --+ oo,
4. We square the equation II I ( 1/f1 + 1/f2) 112 i = 1, 2, to deduce the result. 5.
~). v~t (~2 6t
=
ll1/f1
+ 1/f2ll and use the fact that II I (1/Fi) 112 =
111/Fi II for
The question permits us to use the integrating factor ef3t to give, formally, ef3t Xt
=
__s ds = ef3t Wt - fJ lot ef3s Ws ds lootef3s dW ds o
on integrating by parts. This is the required result, and substitution verifies that it satisfies the given equation. 6.
Find a sequence cp = (cp(n)) of predictable step functions such that llc/J(n)
-1/f II
--+ 0 as n --+ oo.
By the argument before equation (13.8.9), /(cp(n)) ~ /(1/f) as n --+ oo. By Lemma (13.8.4), II/ (cp(n)) 112 = llc/J(n) II, and the claim follows.
13.9 Solutions. Ito's formula 1. The process Z is continuous and adapted with Zo = 0. We have by Theorem (13.8.11) that IE(Zt - Zs I :Fs) = 0, and by Exercise (13.8.6) that
The first claim follows by the Levy characterization of a Wiener process (12. 7.1 0).
417
[13.9.2]-[13.10.2]
Diffusion processes
Solutions
We have inn dimensions that R2 = XI +X~+···+ X~, and the same argument yields that (Xi/ R) dXi is a Wiener process. By Example (13.9.7) and the above,
Zt = ~i
Jd
d(R 2) = 2
Ln Xi dXi + n dt = 2R Ln X· dXi + n dt = 2R dW + n dt. ___!_
i=l R
i=I
2.
Applying Ito's formula (13.9.4) to Y1 = W 14 we obtain dY1 = 4W? dW1 + 6W? dt. Hence,
3.
Apply Ito's formula (13.9.4) to obtain dY1 = W 1 dt + t dW1 • Cf. Exercise (13.8.1).
4.
Note that X 1 =cos Wand X2 =sin W. By Ito's formula (13.9.4), dY = d(Xr + iXz) = dXr + i dXz = d(cos W) + i d(sin W)
= -sin W dW- ~cos W dt + i cos W dW - ~sin W dt. 5.
We apply Ito's formula to obtain:
(a) (1
+ t) dX
(c)
+ dW, + -v'l=X2 dW,
= -X dt
(b) dX = -~X dt
d(~) = -~ (~) dt + (b~a -~b) (~) dW. 13.10 Solutions. Option pricing
1.
(a) We have that E((ae 2 - K)+) = { 00
(aez- K)
1
00
=
~ exp (- (z- ;) 2 )
v2nr2
lrog(Kfa)
I 2
e-zY (aeY+
a
1
= aeY+z'
21 a
$
00
I 2 e-z(y-r)
~
v2n
dz
2r
z- y log(K/a)- y where y = - - , a= - - - - r
r
dy- K
= aeY+ir 2
(b) We have that ST = ae 2 where a = S1 and, under the relevant conditional Q-distribution, Z is normal with mean y = (r - ~a 2 )(T - t) and variance r 2 = a 2(T - t). The claim now follows by the result of part (a). 2. (a) Set ~(t, S) = ~(t, S1 ) and 1/r(t, S) = 1/r(t, S1 ), in the natural notation. By Theorem (13.10.15), we have 1/rx = 1fr1 = 0, whence 1/r(t, x) = c for all t, x, and some constant c. (b) Recall that dS = MS dt +aS dW. Now, d(~S + 1frer 1 ) = d(S 2 + 1frer 1 ) = (aS) 2 dt + 2S dS + ert dlfr + lfrrert dt,
418
Solutions [13.10.3]-[13.10.5]
Option pricing
by Example (13.9.7). By equation (13.10.4), the portfolio is self-financing if this equals S dS 1/frert dt, and thus we arrive at the SDE ert dl{! = -S dS- a 2 S 2 dt, whence '''(t S) --lot e-ru Su dSu - a2 lot e-ru s2u du o/ ' 0
(c) Note first that Zt
=
Jd Su du satisfies dZt
+
0
0
=
St dt. By Example (13.9.8), d(StZt)
=
Zt dSt
+
s[ dt, whence d(~ S + 1/fert)
= Zt dSt +sf dt + ert dl{! + rert dt.
Using equation (13.10.4), the portfolio is self-financing if this equals Zt dSt require that ert dl{! = -S[ dt, which is to say that
+ 1/frert dt, and thus we
3. We need to check equation (13.10.4) remembering that dMt = 0. Each of these portfolios is self-financing. (a) This case is obvious. (b)d(~S + 1/J)
= d(2S 2 - s2 - t) = 2SdS +dt -dt = ~dS. (c) d(~S + 1/1) = -S- tdS + S = ~ dS. (d) Recalling Example (13.9.8), we have that
4. The time of exercise of an American call option must be a stopping time for the filtration (:Ft). The value of the option, if exercised at the stopping time r, is V, = (S, - K)+, and it follows by the usual argument that the value at time 0 of the option exercised at r is E
where N is an N(O, 1) random variable. It is immediate that this is increasing in So and r and is decreasing in K. To show monotonicity in T, we argue as follows. Let T1 < T2 and consider the European option with exercise date T2. In the corresponding American option we are allowed to exercise the option at the earlier time T1 . By Exercise (13.10.4), it is never better to stop earlier than T2, and the claim follows. Monotonicity in a may be shown by differentiation.
419
[13.11.1]-[13.12.2]
Solutions
Diffusion processes
13.11 Solutions. Passage probabilities and potentials 1. LetHbeaclosedspherewithradius R (> lwl), and define PR(r) = Then PR satisfies Laplace's equation in JRd, and hence
!!._ dr
lP'( G before H
II W(O)I = r ).
(rd-ldPR) =O dr
since PR is spherically symmetric. Solve subject to the boundary equations PR(E) = 1, PR(R) = 0, to obtain 2-d - R2-d r d-2 as R-+ oo. PR(r) = 2-d R2-d -+ (E/r) E
-
2. The electrical resistance Rn between 0 and the set /':,.n is no smaller than the resistance obtained by, for every i = 1, 2, ... , 'shorting out' all vertices in the set /':,.i. This new network amounts to a linear chain of resistances in series, points labelled /':,.i and /':,.i+ 1 being joined by a resistance if size Ni-l . It follows that 00 1 R(G)= lim Rn::::L-· n->-oo i=O Ni By Theorem (13.11.18), the walk is persistent if "Ei Ni-l = oo. 3. Thinking of G as an electrical network, one may obtain the network H by replacing the resistance of every edge e lying in G but not in H by oo. Let 0 be a vertex of H. By a well known fact in the theory of electrical networks, R(H) :::: R(G), and the result follows by Theorem (13.11.19).
13.12 Solutions to problems (a) T(t) =a W(t ja 2 ) has continuous sample paths with stationary independent increments, since W has these properties. Also T(t)ja is N(O, tja 2 ), whence T(t) is N(O, t). (b) As for part (a). (c) Certainly V has continuous sample paths on (0, oo). For continuity at 0 it suffices to prove that t W(t- 1 ) -+ 0 a.s. as t ,), 0; this was done in the solution to Exercise (13.6.4). If (u,v), (s,t) are disjoint time-intervals, then so are (v- 1 ,u- 1), (t- 1 ,s- 1); since W has independent increments, so has V. Finally, 1.
is N(O, {3) if s, t > 0, where
f3 =
_!!__ +s2 s+t
(!- _1_) s
s+t
= t.
2. Certainly W is Gaussian with continuous sample paths and zero means, and it is therefore sufficient to prove that cov(W (s), W(t)) = min{s, t}. Now, if s .:::; t,
as required.
420
Solutions [13.12.3]-[13.12.4]
Problems
Ifu(s) = s, v(t) = 1- t, then r(t) this case X(t) = (1- t)W(t/(1- t)). 3.
= tj(l- t), and r- 1(w) = wj(l +
w) forO::; w < oo. In
Certainly U is Gaussian with zero means, and U (0) = 0. Now, with s1 = e 2f3t - 1, E{U(t +h) I U(t)
= u} = e-f3(t+h)E{W(sr+h) I W(sr) = uef3 1 } = ue-f3(t+h)ef3t = u- f3uh + o(h),
whence the instantaneous mean of U is a(t, u) therefore E{U(t
+ h) 2
1
=
-f3u. Secondly, St+h
= s1 +
2{3e 2f3 1h + o(h), and
= u} = e- 2f3(t+h)E{W(s 1+h) 2 W(s 1 ) = uef3 1 } = e-2f3(t+h) (u2e2f3t + 2f3e2f3t h + o(h)) = u 2 - 2f3h(u 2 - 1) + o(h).
U(t)
1
It follows that
E{IU(t +h)- U(t)1 2 l U(t)
=
u}
= u2 - 2f3h(u 2 = 2f3h + o(h),
1)- 2u(u- f3uh) + u 2 + o(h)
and the instantaneous variance is b(t, u) = 2{3.
4.
Bartlett's equation (see Exercise (13.3.4)) for M(t, B)= E(eeV(t)) is
aM at
aM aB
1 2 2
-=-{38-+-a B M 2
with boundary condition M (B, 0) = e 8 u. Solve this equation (as in the exercise given) to obtain
the moment generating function of the given normal distribution. Now M (t, B) --+ exp{ ~B 2 a 2 j (2{3)} as t --+ oo, whence by the continuity theorem V (t) converges in distribution to the N (0, ~a 2 j fJ) distribution. If V (0) has this limit distribution, then so does V (t) for all t. Therefore the sequence (V Ct1), ... , V(tn)) has the same joint distribution as (V(t1 +h), ... , V(tn +h)) for all h, t1; ... , tn, whenever V (0) has this normal distribution. In the stationary case, E(V(t)) = 0 and, for s ::; t, cov(V(s), V(t))
= E{V(s)E(V(t) I V(s))} = E{V(s) 2 e-f3(t-s)} = c(O)e-f31t-sl
where c(O) = var(V (s)); we have used the first part here. This is the autocovariance function of a stationary Gaussian Markov process (see Example (9.6.10)). Since all such processes have autocovariance functions of this form (i.e., for some choice of fJ), all such processes are stationary Omstein-Uhlenbeck processes. The autocorrelation function is p(s) = e-f31sl, which is the characteristic function of the Cauchy density function 1 X ER f(x) = fJn{l + (x/{3)2},
421
[13.12.5]-[13.12.7] 5.
Diffusion processes
Solutions
Bartlett's equation (see Exercise (13.3.2)) forM is
aM at
-
aM + -{JB 1 2 aM ae 2 -ae '
= aB-
subject to M(O, B)= e8d. The characteristics satisfy
The solution isM= g(Beat j(a The solution follows as given. By elementary calculations,
dM
dt
0
1
+ ifJB))
E(D(t)) =
2d() 2aB
where g is a function satisfying g(Bj(a
aM I = ae 8=0
a2 M E(D(t)2) = --2
+ f3() 2 ·
J
ae 8=0
+ ifJB))
= e 8d.
deat,
fJd = -eat (eat - 1)
+ d2e2at'
a
whence var(D(t)) = (fJdja)eat (eat - 1). Finally IP'(D(t) = 0) =
lim
8-->--oo
M(t, B)= exp {
2adeat
t
}
{3(1 - ea )
which converges to e- 2adffJ as t--+ oo.
6.
The equilibrium density function g(y) satisfies the (reduced) forward equation d --(ag) dy
where a(y) are
=
-fJy and b(y)
2
1 d + --(bg) = 2 dy2
0
= a 2 are the instantaneous mean and variance.
The boundary conditions
y = -c,d. Integrate (*) from -c to y, using the boundary conditions, to obtain
-c
~
y
~d.
Integrate again to obtain g(y) = Ae-f3Y 2fa 2 . The constant A is given by the fact that f~c g(y) dy = 1.
7.
First we show that the series converges uniformly (along a subsequence), implying that the limit exists and is a continuous function of t. Set n-1 . (k ) ~ sm t Zmn(t) = ~ -k-Xk>
Mmn = sup{IZmn(t)l: 0 ~ t ~
7r }.
k=m
We have that M2 < mn -
su
n-1 eikt
~ -X P l~ k k
O::or:::,:rr k=m
12
n-1 x2 < ~ ___k_ ~ k2
-
k=m
422
+2
n-m-11n-1-1 X·X· I 1 J+l ~ ~ ~ ~ · ( · + l) · 1=1
J=m 1 1
Solutions
Problems
[13.12.8]-[13.12.8]
The mean value of the final term is, by the Cauchy-Schwarz inequality, no larger than n-m-1
2
1
n-l-1
L:
"
+ [)2
L.J j2(j }=m
1=1
< 2(n - m) -
F.t--
m m4 ·
--
Combine this with (*) to obtain
It follows that
implying that .2:::~ 1 M2n-I 2n < oo a.s. Therefore the series which defines W converges uniformly with probability 1 (along a ;ubsequence), and hence W has (a.s.) continuous sample paths. Certainly W is a Gaussian process since W(t) is the sum of normal variables (see Problem (7.11.19)). Furthermore E(W(t)) = 0, and st
cov ( W(s), W(t) ) = -
2 ~ sin(ks) sin(kt)
+- L.J
T(
T(
2
k
k=1
since the Xi are independent with zero means and unit variances. It is an exercise in Fourier analysis to deduce that cov(W(s), W(t)) = min{s, t}.
8.
We wish to find a solution g(t, y) to the equation
IYI <
b,
satisfying the boundary conditions g(O, y) = 8yo
if IYI :S b,
g(t, y) =
0 if IYI =b.
Let g(t, y I d) be the N(d, t) density function, and note that g(-, · I d) satisfies (*) for any 'source' d. Let 00
L:
g(t, y) =
c-1)kg(t, y 12kb),
k=-00
a series which converges absolutely and is differentiable term by term. Since each summand satisfies (*),so does the sum. Now g(O, y) is a combination of Dirac delta functions, one at each multiple of2b. Only one such multiple lies in [-b, b], and hence g(y, 0) = 8ao· Also, setting y = b, the contributions from the sources at -2(k- 1)b and 2kb cancel, so that g(t, b) = 0. Similarly g(t, -b) = 0, and therefore g is the required solution. Here is an alternative method. Look for the solution to(*) of the form e-l.nt sin{inn(y + b)jb}; such a sine function vanishes when IYI =b. Substitute into(*) to obtain An = n 2 n 2 /(8b 2 ). A linear combination of such functions has the form 00 g(t,y) =Lane
-Ant .
sm
n=1
423
(
nn(y +b) ) . 2b
[13.12.9]-[13.12.11]
Solutions
Diffusion processes
We choose the constants an such that g(O, y) = 8yo for IYI < b. With the aid of a little Fourier analysis, one finds that an = b- 1 sin(~mr). Finally, the required probability equals the probability that wa has been absorbed by time t, a probability expressible as 1- f~b fa(t, y) dy. Using the second expression for this yields
r,
9. Recall that U(t) = e- 2mD(t) is a martingale. LetT be the time of absorption, and assume that the conditions of the optional stopping theorem are satisfied. Then E(U(O)) = E(U(T)), which is to say that 1 = e2ma Pa + e-2mb(l - Pa).
10. (a) We may assume that a, b > 0. With Pt(b)
= lP'(W(t)
> b, F(O, t) I W(O) =a),
we have by the reflection principle that Pt(b) = lP'(W(t) > b I W(O) =a) -lP'(W(t) < -b I W(O) =a) = lP'(b- a< W(t) < b +a I W(O) =
o),
giving that apr (b) ---az;=
f(t, b +a)- f(t, b- a)
where f(t, x) is the N(O, t) density function. Now, using conditional probabilities, lP'(F(O, t) I W(O) =a, W(t) =b)=-
1 apr(b) = 1 _ e-2abft. f(t, b- a) ab
(b) We know that
2
lP'(F(s, t)) = 1 - - cos- 1 { T(
VS/t} = -2 sin- 1 { VS/t} T(
if 0 < s < t. The claim follows since F (to, t2) s; F (t1, t2). (c) Remember that sinx = x + o(x) as x 0. Take the limit in part (b) as to
+
+0 to obtain ,fiJ!i2.
11. Let M(t) = sup{W(s): 0:::; s:::; t} and recall that M(t) has the same distribution as IW(t)l. By symmetry, lP'( sup IW(s)l::::
w) :::; 2lP'(M(t):::: w)
= 2lP'(IW(t)l::::
w).
O~s~t
By Chebyshov's inequality,
Fix E > 0, and let An(E) = {I W(s)l/s > E for somes satisfying 2n- 1 < s :::; 2n }. Note that
424
Solutions [13.12.12]-[13.12.13]
Problems
for all large n, and also
Therefore L:n lP'(An(E)) < oo, implying by the Borel-Cantelli lemma that (a.s.) only finitely many of the An(E) occur. Therefore t- 1W(t)--+ 0 a.s. as t--+ oo. Compare with the solution to the relevant part of Exercise (13.6.4).
12. We require the solution to Laplace's equation p(w)
=
v2 p
= 0, subject to the boundary condition
0 ifwEH, { 1 'f G. 1 W E
Look for a solution in polar coordinates of the form 00
p(r, e) =
L rn {an sin(ne) + bn cos(ne)}. n=O
Certainly combinations having this form satisfy Laplace's equation, and the boundary condition gives that 00
H(e) = bo
+L
{an sin(ne)
+ bn cos(ne) },
1e1
< n,
n=1 where H(e) = {
~
if - n <
e < o,
ifO
The collection {sin(me), cos( me) : m 2: 0} are orthogonal over ( -n, n). Multiply through(*) by sin(me) and integrate over ( -n, n) to obtain nam = {1 - cos(nm)}/m, and similarly bo = ~ and bm = 0 for m 2: 1.
13. The joint density function of two independent N (0, t) random variables is (2n t) - 1 exp{- (x 2 + y2)j(2t)}. Since this function is unchanged by rotations of the plane, it follows that the two coordinates of the particle's position are independent Wiener processes, regardless of the orientation of the coordinate system. We may thus assume that lis the line x = d for some fixed positive d. The particle is bound to visit the line l sooner or later, since lP'(W1 (t) < d for all t) = 0. The first-passage timeT has density function t > 0. Conditional on {T = t}, D = W2(T) is N(O, t). Therefore the density function of Dis fv(u) =
loo
oo
fDIT(u I t)fT(t) dt =
looo --e-(u d +d )/( 2 2
2
o 2m 2
t) dt
=
d n(u 2
giving that DId has the Cauchy distribution. The angle E> = PoR satisfies e = tan- 1(D/d), whence lP'(E>.::; e)= lP'(D.::; dtane) =
425
1
e
2 + -;·
1e1
< ~n.
+ d2)
,
u E ffi:.,
[13.12.14]-[13.12.18]
Solutions
Diffusion processes
14. By an extension of Ito's formula to functions of two Wiener processes, U = u(Wr, W2 ) and V = v(Wr, W2) satisfy dU = Ux dWr dV = Vx dWr
+ uy dW2 + J-Cuxx + Uyy) dt, + vy dW2 + J-Cvxx + Vyy) dt,
where Ux, Vyy, etc, denote partial derivatives of u and v. Since ¢ is analytic, u and v satisfy the Cauchy-Riemann equations Ux = vy, uy = -vx, whence u and v are harmonic in that Uxx + uyy = Vxx + Vyy = 0. Therefore,
The matrix (
~xUy
uy) is an orthogonal rotation oflH:2 when u~ +u~ = 1. Since the joint distribution Ux
of the pair (Wr, W2) is invariant under such rotations, the claim follows. 15. One method of solution uses the fact that the reversed Wiener process {W (t- s)- W (t) : 0 ::: s ::: t} has the same distribution as {W(s) : 0 ::: s ::: t}. Thus M(t)- W (t) = maxo~s~dW(s)- W(t)} has the same distribution as maxo~u~dW(u)- W(O)} = M(t). Alternatively, by the reflection principle, lP'(M(t) 2':: x, W(t)::: y) = lP'(W(t) 2':: 2x- y)
for x 2':: max{O, y}.
By differentiation, the pair M(t), W(t) has joint density function -2¢' (2x - y) for y ::: x, x 2':: 0, where¢ is the density function of the N(O, t) distribution. Hence M(t) and M(t)- W(t) have the joint density function -2¢'(x + y). Since this function is symmetric in its arguments, M(t) and M(t)- W(t) have the same marginal distribution. 16. The Lebesgue measure A(Z) is given by
A(Z)
= fooo l{W(t)=u) du,
whence by Fubini's theorem (cf. equation (5.6.13)), JE:(A(Z))
= fooo lP'(W(t) = u) dt = 0.
17. Let 0
= maxx 9 ~y W(s).
M(c, d)- M(a, b)= max {W(s)- W(c)} c~s~d
Then
+ {W(c)- W(b)}-
max {W(s)- W(b) }.
a~s~b
Since the three terms on the right are independent and continuous random variables, it follows that lP'((M(c, d) = M(a, b)) = 0. Since there are only countably many rationals, we deduce that lP'( (M(c, d) = M(a, b) for all rationals a < b < c
18. The result is easily seen by exhaustion to be true when n = 1. Suppose it is true for all m ::: n - 1 where n 2':: 2. (i) If sn ::: 0, then (whatever the final term of the permutation) the number of positive partial sums and the position of the first maximum depend only on the remaining n - 1 terms. Equality follows by the induction hypothesis. (ii) If Sn > 0, then n
Ar =
L Ar-1 (k), k=l
426
Solutions
Problems
[13.12.19]-[13.12.20]
where Ar-1 (k) is the number of permutations with Xk in the final place, for which exactly r - 1 of the first n- 1 terms are strictly positive. Consider a permutation rc = (xi 1 , xi 2 , ... , Xin-I, Xk) with Xk in the final place, and move the positionofxk to obtain thenewpermutationrc' = (xk. xi 1 , Xi 2 , •.• , Xin-l ). The first appearance of the maximum in rc' is at its rth place if and only if the first maximum of the reduced permutation (xi 1 , Xi 2 , ... , X in- I) is at its (r - l)th place. [Note that r = 0 is impossible since sn > 0.] It follows that n
Br =
L
Br-1 (k), k=1 where Br-1 (k) is the number of permutations with Xk in the final place, for which the first appearance of the maximum is at the (r - 1)th place. By the induction hypothesis, Ar-1 (k) = Br-1 (k), since these quantities depend on the n - 1 terms excluding Xk. The result follows.
2::f=
19. Suppose that Sm = 1 Xj, 0 ::": m ::": n, are the partial sums of n independent identically distributed random variables Xj. Let An be the number of strictly positive partial sums, and Rn the index of the first appearance of the value of the maximal partial sum. Each of the n! permutations of (X 1, X2, ... , Xn) has the same joint distribution. Consider the kth permutation, and let h be the indicator function of the event that exactly r partial sums are positive, and let h be the indicator function that the first appearance ofthe maximum is at the rth place. Then, using Problem (13.12.18), 1 n! 1 n! lP'(An = r) = - LE(h) = - LE(h) = lP'(Rn = r). n! k=1 n! k=1 We apply this with Xj
= W(jtjn)- W((j-
l)t/n), so that Sm
l::j I{W(jtfn)>O) has the same distribution as
Rn = min{k 2':: 0: W(kt/n) =
= W(mtjn).
Thus An
m;:tx W(jtjn) }. O:'OJ:'On
By Problem (13.12.17), Rn ~ R as n---+ oo. By Problem (13.12.16), the time spent by W at zero is a.s. a null set, whence An ~ A. Hence A and R have the same distribution. We argue as follows to obtain that that L and R have the same distribution. Making repeated use of Theorem (13.4.6) and the symmetry of W, lP'(L < x) =
lP'c~~~t W(s)
<
0) + lP'cl~~t W(s) > 0)
= 2lP'( sup {W(s)- W(x)} < -W(x)) = 2lP'(IW(t)- W(x)l < W(x)) x:'Os:'Ot
= lP'(IW(t)- W(x)l < IW(x)l)
= lP'(
sup {W(s)- W(x)} < X:'OS:'Ot
sup {W(s)- W(x)})
= lP'(R
::": x).
0:'0S:'OX
Finally, by Problem (13.12.15) and the circular symmetry of the joint density distribution of two independent N(O, 1) variables U, V, lP'(IW(t)- W(x)l < IW(x)l) = lP'((t- x)V 2 ::::: xU 2 ) = lP' (
20. Let Tx = {
inf {t ::": 1 : W (t)
= x}
1
2
:::::
~) = ~ sin- 1 ~. t Vt
if this set is non-empty, otherwise,
427
V
u2 +V 2
Jr
[13.12.21]-[13.12.24]
Diffusion processes
Solutions
and similarly Vx = sup{t .::; 1 : W(t) = x}, with Vx = 1 if W(t) "# x for all t E [0, 1]. Recall that Uo and Vo have an arc sine distribution as in Problem (13.12.19). On the event {Ux < 1}, we may write (using the re-scaling property of W) Ux
=
Tx
+ (1
- Tx)Uo,
=
Vx
Tx
+ (1 -
Tx)Vo,
where Uo and Yo are independent of Ux and Vx, and have the above arc sine distribution. Hence Ux and Vx have the same distribution. Now Tx has the first passage distribution of Theorem (13.4.5), whence
f
- (r r/J)
Tx,Uo
'
x2)} {
= { -x- exp ( - v'2nr3 2r
1 } n./r/J(1- rp) ·
Therefore,
and 1 fux(u)= lo u frxux(t,u)du= exp ( -x2) - , o ' n.ju(1-u) 2u
O
21. Note that Vis a martingale, by Theorem (13.8.11). Fix t and let 1/fs = sign(Ws), 0.::; s .::; t. We have that 111/fll = v'f, implying by Exercise (13.8.6) that E(V?) = ll/(1/f)ll~ = t. By a similar calculation, E(V? 1 :Fs) = v? + t - s for 0 .::; s .::; t. That is to say, v? - t defines a martingale, and the result follows by the Levy characterization theorem of Example (12.7.10). 22. The mean cost per unit time is f.l(T) =
~{ R +CloT lP'(IW(t)l 2: a) dt} = ~{ R + 2C loT (1-
Differentiate to obtain that 11' (T) = 0 if R = 2C {loT
where we have integrated by parts.
23. Consider the portfolio with Ht, S1) units of stock and 1/f(t, S1) units of bond, having total value w(t, S1) = x~(t. x) + ert 1/f(t, S1). By assumption, (1- y)x~(t, x) = yer 1 1jf(t, x). Differentiate this equation with respect to x and substitute from equation (13.10.16) to obtain the differential equation (1- y)~ + x~x = 0, with solution ~(t, x) = h(t)xY-I, for some function h(t). We substitute this, together with(*), into equation (13.10.17) to obtain that h'- h(l- y)(~yo- 2
+ r) =
0.
It follows that h(t) = A exp{(1 - y)( ~yo- 2 + r)t}, where A is an absolute constant to be determined according to the size of the initial investment. Finally, w(t, x) = y- 1 x~(t. x) = y- 1h(t)xY.
24. Using Ito's formula (13.9.4), the drift term in the SDE for U1 is (-ut(T- t, W)
+ ~u22(T- t, W)) dt,
where u1 and u22 denote partial derivatives of u. The drift function is identically zero if and only if -
U[ -
1
zU22.
428
Bibliography
A man will turn over half a library to make one book.
Samuel Johnson
Abramowitz, M. and Stegun, I. A. ( 1965). Handbook ofmathematicalfunctions with formulas, graphs and mathematical tables. Dover, New York. Billingsley, P. (1995). Probability and measure (3rd edn). Wiley, New York. Breiman, L. (1968). Probability. Addison-Wesley, Reading, MA, reprinted by SIAM, 1992. Chung, K. L. (1974). A course in probability theory (2nd edn). Academic Press, New York. Cox, D. R. and Miller, H. D. (1965). The theory of stochastic processes. Chapman and Hall, London. Doob, J. L. (1953). Stochastic processes. Wiley, New York. Feller, W. (1968). An introduction to probability theory and its applications, Vol. 1 (3rd edn). Wiley, New York. Feller, W. (1971). An introduction to probability theory and its applications, Vol. 2 (2nd edn). Wiley, New York. Grimmett, G. R. and Stirzaker, D. R. (2001). Probability and random processes, (3rd edn). Oxford University Press, Oxford. Grimmett, G. R. and Welsh, D. J. A. (1986). Probability, an introduction. Clarendon Press, Oxford. Hall, M. (1983). Combinatorial theory (2nd edn). Wiley, New York. Harris, T. E. (1963). The theory of branching processes. Springer, Berlin. Karlin, S. and Taylor, H. M. (1975). A first course in stochastic processes (2nd edn). Academic Press, New York. Karlin, S. and Taylor, H. M. (1981). A second course in stochastic processes. Academic Press, New York. Laha, R. G. and Rohatgi, V. K. (1979). Probability theory. Wiley, New York. Loeve, M. (1977). Probability theory, Vol. 1 (4th edn). Springer, Berlin. Loeve, M. (1978). Probability theory, Vol. 2 (4th edn). Springer, Berlin. Moran, P. A. P. (1968). An introduction to probability theory. Clarendon Press, Oxford. Stirzaker, D. R. (1994). Elementary probability. Cambridge University Press, Cambridge. Stirzaker, D. R. (1999). Probability and random variables. Cambridge University Press, Cambridge. Williams, D. (1991). Probability with martingales. Cambridge University Press, Cambridge.
429
Index
Abbreviations used in this index: c.f. characteristic function; distn distribution; eqn equation; fn function; m.g.f. moment generating function; p.g.f. probability generating function; pr. process; r.v. random variable; r.w. random walk; s.r.w. simple random walk; thm theorem.
A absolute value of s.r. w. 6.1.3 absorbing barriers: s.r.w. 1.7.3, 3.9.1, 5, 3.11.23, 25-26, 12.5.4-5, 7; Wiener pr. 12.9.22-3, 13.12.8-9 absorbing state 6.2.1 adapted process 13.8.6 affine transformation 4.13.11; 4.14.60 age-dependent branching pr. 5.5.1-2; conditional 5.1.2; honest martingale 12.9.2; mean 10.6.13 age, see current life airlines 1.8.39, 2. 7. 7 alarm clock 6.15.21 algorithm 3.11.33, 4.14.63, 6.14.2 aliasing method 4.11.6 alternating renewal pr. 10.5.2, 10.6.14 American call option 13.10.4 analytic fn 13.12.14 ancestors in common 5.4.2 anomalous numbers 3.6.7 Anscombe's theorem 7.11.28 antithetic variable 4.11.11 ants 6.15.41 arbitrage 3.3.7, 6.6.3 Arbuthnot, J. 3.11.22 arc sine distn 4.11.13; sample from 4.11.13 arc sine law density 4.1.1 arc sine laws for r.w.: maxima 3.11.28; sojourns 5.3.5; visits 3.10.3
arc sine laws for Wiener pr. 13.4.3, 13.12.10, 13.12.19 Archimedes's theorem 4.11.32 arithmetic r.v. 5.9.4 attraction 1.8.29 autocorrelation function 9.3.3, 9.7.5, 8 autocovariance function 9.3.3, 9.5.2, 9.7.6, 19-20, 22 autoregressive sequence 8.7.2, 9.1.1, 9.2.1, 9.7.3 average, see moving average
B babies 5.10.2 backward martingale 12.7.3 bagged balls 7.11.27, 12.9.13-14 balance equations 11.7.13 Bandrika 1.8.35-36, 4.2.3 bankruptcy, see gambler's ruin Barker's algorithm 6.14.2 barriers: absorbing/retaining in r.w. 1.7.3, 3.9.1-2, 3.11.23, 25-26; hitting by Wiener pr. 13.4.2 Bartlett: equation 13.3.2-4; theorem 8.7.6, 11.7.1 batch service 11.8.4 baulking 8.4.4, 11.8.2, 19 Bayes's formula 1.8.14, 1.8.36 bears 6.13.1, 10.6.19 Benford's distn 3.6.7 Berkson's fallacy 3.11.37 Bernoulli: Daniel 3.3.4, 3.4.3-4; Nicholas 3.3.4
430
Bernoulli: model 6.15.36; renewal 8.7.3; shift 9.17.14; sum ofr.v.s 3.11.14, 35 Bertrand's paradox 4.14.8 Bessel: function 5.8.5, 11.8.5, 11.8.16; B. pr. 12.9.23, 13.3.5-6, 13.9.1 best predictor 7.9.1; linear 7.9.3, 9.2.1-2, 9.7.1, 3 beta fn 4.4.2, 4.10.6 beta distn: b.-binomial 4.6.5; sample from 4.11.4-5 betting scheme 6.6.3 binary: expansion 9 .1.2; fission 5.5.1 binary tree 5.12.38; r.w. on 6.4.7 binomial r.v. 2.1.3; sum of 3.11.8, 11 birth process 6.8.6; dishonest 6.8.7; forward eqns 6.8.4; divergent 6.8.7; with immigration 6.8.5; non-homogeneous 6.15.24; see also simple birth birth-death process: coupled 6.15.46; extinction 6.11.3, 5, 6.15.25, 12.9.10; honest 6.15.26; immigration-death 6.11.3, 6.13.18, 28; jump chain 6.11.1; martingale 12.9.10; queue 8.7.4; reversible 6.5.1, 6.15.16; symmetric 6.15.27; see also simple birth-death birthdays 1.8 .30 bivariate: Markov chain 6.15.4; negative binomial distn 5.12.16; p.g.f. 5.1.3 bivariate normal distn 4.7.5-6, 12, 4.9.4-5, 4.14.13, 16, 7.9.2,
Index 7.11.19; c.f. 5.8.11; positive
part 4.7.5, 4.8.8, 5.9.8 Black-Scholes: model 13.12.23; value 13.10.5 Bonferroni's inequality 1.8.37 books 2.7.15 Boole's inequalities 1.8.11 Borel: normal number theorem 9.7.14; paradox 4.6.1 Borel-Cantelli lemmas 7.6.1, 13.12.11 bounded convergence 12.1.5 bow tie 6.4.11 Box-Muller normals 4.11.7 branching process: age-dependent 5.5.1-2, 10.6.13; ancestors 5.4.2; conditioned 5.12.21, 6.7.1-4; convergence, 12.9.8; correlation 5.4.1; critical 7.10.1; extinction 5.4.3; geometric 5.4.3, 5.4.6; imbedded in queue 11.3.2, 11.7.5, 11; with immigration 5.4.5, 7.7.2; inequality 5.12.12; martingale 12.1.3, 9, 12.9.1-2, 8; maximum of 12.9.20; moments 5.4.1; p.g.f. 5.4.4; supercritical 6.7.2; total population 5.12.11; variance 5.12.9; visits 5.4.6 bridge 1.8.32 Brownian bridge 9.7.22, 13.6.2-5; autocovariance 9.7.22, 13.6.2; zeros of 13.6.5 Brownian motion; geometric 13.3.9; tied-down, see Brownian bridge Buffon: cross 4.5.3; needle 4.5.2, 4.14.31-32; noodle 4.14.31 busy period 6.12.1; in G/G/1 11.5.1; in M/G/1 11.3.3; in M/M/1 11.3.2, 11.8.5; in M!M/oo 11.8.9
c cake, hot 3.11.32 call option: American 13.10.4; European 13.10.4-5 Campbell-Hardy theorem 6.13.2 capture-recapture 3.5.4 car, parking 4.14.30 cards 1.7.2, 5, 1.8.33 Carroll, Lewis 1.4.4 casino 3.9.6, 7.7.4, 12.9.16 Cauchy convergence 7.3.1; in m.s. 7.11.11 Cauchy distn 4.4.4; maximum 7 .11.14; moments 4.4.4;
sample from 4.11.9; sum 4.8.2, 5.11.4, 5.12.24-25 Cauchy-Schwarz inequality 4.14.27 central limit theorem 5.10.1, 3, 9, 5.12.33, 40, 7.11.26, 10.6.3 characteristic function 5.12.26-31; bivariate normal 5.8.11; continuity theorem 5.12.39; exponential distn 5.8.8; extreme-value distn 5.12.27; first passage distn 5.10.7-8; joint 5.12.30; law of large numbers 7.11.15; multinormal distn 5.8.6; tails 5.7.6; weak law 7.11.15 Chebyshov's inequality, one-sided 7.11.9 cherries 1.8.22 chess 6.6.6-7 chimeras 3.11.36 chi-squared distn: non-central 5.7.7; sum 4.10.1, 4.14.12 Cholesky decomposition 4.14.62 chromatic number 12.2.2 coins: double 1.4.3; fair 1.3.2; first head 1.3.2, 1.8.2, 2.7.1; patterns 1.3.2, 5.2.6, 5.12.2, 10.6.17, 12.9.16; transitive 2.7.16; see Poisson flips colouring: graph 12.2.2; sphere 1.8.28; theorem 6.15.39 competition lemma 6.13.8 complete convergence 7.3.7 complex-valued process 9.7.8 compound: Poisson pr. 6.15.21; Poisson distn 3.8.6, 5.12.13 compounding 5.2.3, 5.2.8 computer queue 6.9.3 concave fn 6.15.37 conditional: birth-death pr. 6.11.4-5; branching pr. 5.12.21, 6.7.1-4; convergence 13.8.3; correlation 9.7.21; entropy 6.15.45; expectation 3.7.2-3, 4.6.2, 4.14.13, 7.9.4; independence 1.5.5; probability 1.8.9; s.r.w. 3.9.2-3; variance 3.7.4, 4.6.7; Wiener pr. 8.5.2, 9.7.21, 13.6.1; see also regression continuity of: distn fns 2. 7.1 0; marginals 4.5.1; probability measures 1.8.16, 1.8.18; transition probabilities 6.15.14 continuity theorem 5.12.35, 39
431
continuous r.v.: independence 4.5.5, 4.14.6; limits of discrete r.v.s 2.3.1 convergence: bounded 12.1.5; Cauchy 7.3.1, 7.11.11; complete 7.3.7; conditional 13.8.3; in distn 7.2.4, 7.11.8, 16, 24; dominated 5.6.3, 7.2.2; event of 7.2.6; martingale 7.8.3, 12.1.5, 12.9.6; Poisson pr. 7.11.5; in probability 7.2.8, 7.11.15; subsequence 7.11.25; in total variation 7 .2.9 convex: fn 5.6.1, 12.1.6-7; rock 4.14.47; shape 4.13.2-3, 4.14.61 corkscrew 8.4.5 Com Flakes 1.3.4, 1.8.13 countable additivity 1.8.18 counters 10.6.6-8, 15 coupling: birth-death pr. 6.15.46; maximal 4.12.4-6, 7.11.16 coupons 3.3.2, 5.2.9, 5.12.34 covariance: matrix 3.11.15, 7.9.3; of Poisson pr. 7 .11.5 Cox process 6.15.22 Cp inequality Cramer-Wold device 7.11.19, 5.8.11 criterion: irreducibility 6.15.15; Kolmogorov's 6.5.2; for persistence 6.4.1 0 Crofton's method 4.13.9 crudely stationary 8.2.3 cube: point in 7.11.22; r.w. on 6.3.4 cumulants 5.7.3-4 cups and saucers 1.3.3 current life 10.5.4; and excess 10.6.9; limit 10.6.4; Markov 10.3.2; Poisson 10.5.4
D dam 6.4.3 dead period 10.6.6-7 death-immigration pr. 6.11.3 decay 5.12.48, 6.4.8 decimal expansion 3.1.4, 7 .11.4 decomposition: Cholesky 4.14.62; Krickeberg 12.9.11 degrees of freedom 5.7.7-8 delayed renewal pr. 10.6.12 de Moivre: martingale 12.1.4, 12.4.6; trial 3.5.1 De Morgan laws 1.2.1 density: arc sine 4.11.13; arc sine law 4.1.1; betsa 4.11.4,
Index
4.14.11, 19, 5.8.3; bivariate normal 4.7.5-6, 12, 4.9.4-5, 4.14.13, 16, 7.9.2, 7.11.19; Cauchy 4.4.4, 4.8.2, 4.10.3, 4.14.4, 16, 5.7.1, 5.11.4, 5.12.19, 24-25, 7.11.14; chi-squared 4.10.1, 4.14.12, 5.7.7; Dirichlet 4.14.58; exponential 4.4.3, 4.5.5, 4.7.2, 4.8.1, 4.10.4, 4.14.4-5, 17-19, 24, 33, 5.12.32, 39, 6.7.1; extreme-value 4.1.1, 4.14.46, 7.11.13; F(r, s) 4.10.2, 4, 5.7.8; first passage 5.10.7-8, 5.12.18-19; Fisher's spherical 4.14.36; gamma 4.14.10-12, 5.8.3, 5.9.3, 5.10.3, 5.12.14, 33; hypoexponential 4.8.4; log-normal 4.4.5, 5.12.43; multinormal 4.9.2, 5.8.6; normal4.9.3, 5, 4.14.1, 5.8.4-6, 5.12.23, 42, 7.11.19; spectral 9.3 .3; standard normal 4.7.5; Student's t 4.10.2-3, 5.7.8; uniform 4.4.3, 4.5.4, 4.6.6, 4.7.1, 3, 4, 4.8.5, 4.11.1, 8, 4.14.4, 15, 19, 20, 23-26, 5.12.32, 7.11.4, 9.1.2, 9.7.5; Weibull 4.4.7, 7.11.13 departure pr. 11.2.7, 11.7.2-4, 11.8.12 derangement 3.4.9 diagonal selection 6.4.5 dice 1.5.2, 3.2.4, 3.3.3, 6.1.2; weighted/loaded 2.7.12, 5.12.36 difference eqns 1.8.20, 3.4.9, 5.2.5 difficult customers 11.7.4 diffusion: absorbing barrier 13.12.8-9; Bessel pr. 12.9.23, 13.3.5-6, 13.9.1; Ehrenfest model 6.5.5, 36; first passage 13.4.2; Ito pr. 13.9.3; models 3.4.4, 6.5.5, 6.15.12, 36; Omstein-Uhlenbeck pr. 13.3.4, 13.7.4-5, 13.12.3-4, 6; osmosis 6.15.36; reflecting barrier 13.5.1, 13.12.6; Wiener pr. 12.7.22-23; zeros 13.4.1, 13.12.10; Chapter 13 passim diffusion approximation to birth-death pr. 13.3 .1 dimer problem 3.11.34 Dirichlet: density 4.14.58; distn 3.11.31 disasters 6.12.2-3, 6.15.28 discontinuous marginal 4.5.1 dishonest birth pr. 6.8.7
distribution: see also density; arc sine 4.11.13; arithmetic 5.9.4; Benford 3.6.7; Bernoulli 3.11.14, 35; beta 4.11.4; beta-binomial 4.6.5; binomial 2.1.3, 3.11.8, 11, 5.12.39; bivariate normal 4.7.5-6, 12; Cauchy 4.4.4; chi-squared 4.10.1; compound 5.2.3; compound Poisson 5.12.13; convergence 7 .2.4; Dirichlet 3.11.31, 4.14.58; empirical 9.7.22; exponential 4.4.3, 5.12.39; F(r, s) 4.10.2-4; extreme-value 4.1.1, 4.14.46, 7.11.13; first passage 5.10.7-8. 5.12.18-19; gamma 4.14.10-12; Gaussian, see normal; geometric 3.1.1, 3.2.2, 3.7.5, 3.11.7, 5.12.34, 39, 6.11.4; hypergeometric 3 .11.1 0-11; hypoexponential 4.8.4; infinitely divisible 5.12.13-14; inverse square 3.1.1; joint 2.5.4; lattice 5.7.5; logarithmic 3 .1.1, 5 .2.3; log-normal 4.4.5, 5.12.43; maximum 4.2.2, 4.14.17; median 2.7.11, 4.3.4, 7.3.11; mixed 2.1.4, 2.3.4, 4.1.3; modified Poisson 3 .1.1; moments 5.11.3; multinomial 3.5.1, 3.6.2; multinormal 4.9.2; negative binomial 3.8.4, 5.2.3, 5.12.4, 16; negative hypergeometric 3.5.4; non-central 5.7.7-8; normal 4.4.6, 8, 4.9.3-5; Poisson 3.1.1, 3.5.2-3, 3.11.6, 4.14.11, 5.2.3, 5.10.3, 5.12.8, 14, 17, 33, 37, 39, 7.11.18; spectral 9.3.2, 4; standard normal 4.7.5; stationary 6.9.11; Student's t 4.10.2-3, 5.7.8; symmetric 3.2.5; tails 5.1.2, 5.6.4, 5.11.3; tilted 5.7.11; trapezoidal 3.8.1; trinormal 4.9.8-9; uniform 2.7.20, 3.7.5, 3.8.1, 5.1.6, 9.7.5; Weibull 4.4.7, 7.11.13; zeta or Zipf 3.11.5 divergent birth pr. 6.8.7 divine providence 3.11.22 Dobrushin's bound and ergodic coefficient 6.14.4 dog-flea model 6.5.5, 6.15.36 dominated convergence 5.6.3, 7.2.2 Doob's Lz inequality 13.7.1 Doob-Kolmogorov inequality 7.8.1-2
432
doubly stochastic: matrix 6.1.12, 6.15.2; Poisson pr. 6.15.22-23 downcrossings inequality 12.3.1 drift 13.3.3, 13.5.1, 13.8.9, 13.12.9 dual queue 11.5.2 duration of play 12.1.4
E Eddington's controversy 1.8.27 editors 6.4.1 eggs 5.12.13 Ehrenfest model 6.5.5, 6.15.36 eigenvector 6.6.1-2, 6.15.7 embarrassment 2.2.1 empires 6.15.10 empirical distn 9.7.22 entrance fee 3.3.4 entropy 7 .5.1; conditional 6.15.45; mutual 3.6.5 epidemic 6.15.32 equilibrium, see stationary equivalence class 7 .1.1 ergodic: coefficient 6.14.4; measure 9.7.11; stationary measure 9.7.11 ergodic theorem: Markov chain 6.15.44, 7.11.32; Markov pr. 7.11.33, 10.5.1; stationary pr. 9.7.10-11, 13 Erlang's loss formula 11.8.19 error 3.7.9; of prediction 9.2.2 estimation 2.2.3, 4.5.3, 4.14.9, 7.11.31 Euler: constant 5.12.27, 6.15.32; product 5.12.34 European call option 13.10.4-5 event: of convergence 7.2.6; exchangeable 7.3.4-5; invariant 9.5.1; sequence 1.8.16; tail 7.3.3 excess life 10.5.4; conditional 10.3.4; and current 10.6.9; limit 10.3.3; Markov 8.3.2, 10.3.2; moments 10.3.3; Poisson 6.8.3, 10.3.1, 10.6.9; reversed 8.3.2; stationary 10.3.3 exchangeability 7.3.4-5 expectation: conditional 3.7.2-3, 4.6.2, 4.14.12, 7.9.4; independent r.v.s 7.2.3; linearity 5.6.2; tail integral 4.3.3, 5; tail sum 3.11.13, 4.14.3 exponential distn: c.f. 5.8.8; holding time 11.2.2; in Poisson pr. 6.8.3; lack-of-memory
Index property 4.14.5; limit in branching pr. 5.6.2, 5.12.21, 6. 7.1; limit of geometric distn 5.12.39; heavy traffic limit 11.6.1; distn of maximum 4.14.18; in Markov pr. 6.8.3, 6.9.9; order statistics 4.14.33; sample from 4.14.48; sum 4.8.1, 4, 4.14.10, 5.12.50, 6.15.42 exponential martingale 13.3.9 exponential smoothing 9.7.2 extinction: of birth-death pr. 6.11.3, 6.15.25, 27, 12.9.10; of branching pr. 6.7.2-3 extreme-value distn 4.1.1, 4.14.46, 5.12.34, 7.11.13; c.f. and mean 5.12.27
F F(r, s) distn 4.10.2, 4;
non-central 5.7.8 fair fee 3.3.4 families 1.5.7, 3.7.8 family, planning 3.11.30 Farkas's theorem 6.6.2 filter 9. 7.2 filtration 12.4.1-2, 7 fingerprinting 3.11.21 finite: Markov chain 6.5.8, 6.6.5, 6.15.43-44; stopping time 12.4.5; waiting room 11.8.1 first passage: c.f. 5.10.7-8; diffusion pr. 13.4.2; distn 5.10.7-8, 5.12.18-19; Markov chain 6.2.1, 6.3.6; Markov pr. 6.9.5-6; mean 6.3.7; m.g.f. 5.12.18; s.r.w. 5.3.8; Wiener pr. 13.4.2 first visit by s.r.w. 3.10.1, 3 Fisher: spherical distn 4.14.36; F.-Tippett-Gumbel distn 4.14.46 FKG inequality 3.11.18, 4.11.11 flip-flop 8.2.1 forest 6.15.30 Fourier: inversion thm 5.9.5; series 9.7.15, 13.12.7 fourth moment strong law 7 .11.6 fractional moments 3.3.5, 4.3.1 function, upper-class 7.6.1 functional eqn 4.14.5, 19
gambling: advice 3.9.4; systems 7.7.4 gamma distn 4.14.10-12, 5.8.3, 5.9.3, 5.10.3, 5.12.14, 33; g. and Poisson 4.14.11; sample from 4.11.3; sum 4.14.11 gamma fn 4.4.1, 5.12.34 gaps: Poisson 8.4.3, 10.1.2; recurrent events 5.12.45; renewal 10.1.2 Gaussian distn, see normal distn Gaussian pr. 9.6.2-4, 13.12.2; Markov 9.6.2; stationary 9.4.3; white noise 13 .8.5 generator 6.9.1 geometric Brownian motion, Wiener pr. 13.3.9 geometric distn 3.1.1, 3.2.2, 3.7.5, 3.11.7, 5.12.34, 39; lack-of-memory property 3.11.7; as limit 6.11.4; sample from 4.11.8; sum 3.8.3-4 goat 1.4.5 graph: colouring 12.2.2; r.w. 6.4.6, 9, 13.11.2-3
H Hajek-Renyi-Chow inequality 12.9.4-5 Hall, Monty 1.4.5 Hastings algorithm 6.14.2 Hawaii 2.7.17 hazard rate 4.1.4, 4.4.7; technique 4.11.10 heat eqn 13.12.24 Heathrow 10.2.1 heavy traffic 11.6.1, 11.7.16 hen, see eggs Hewitt-Savage zero-one law 7.3.4-5 hitting time 6.9.5-6; theorem 3.10.1, 5.3.8 Hoeffding's inequality 12.2.1-2 Holder's inequality 4.14.27 holding time 11.2.2 homogeneous Markov chain 6.1.1 honest birth-death pr. 6.15.26 Hotelling's theorem 4.14.59 house 4.2.1, 6.15.20, 51 hypergeometric distn 3.11.10-11 hypoexponential distn 4.8.4
I
G
idle period 11.5.2, 11.8.9 Galton's paradox 1.5.8 imbedding: jump chain 6.9.11; Markov chain 6.9.12, 6.15.17; gambler's ruin 3.11.25-26, 12.1.4, queues: D/M/1 11.7.16; GIM/1 12.5.8
433
11.4.1, 3; M/G/1 11.3.1, 11.7.4; unsuccessful 6.15.17 immigration: birth-i. 6.8.5; branching 5.4.5, 7.7.2; i.-death 6.11.2, 6.15.18; with disasters 6.12.2-3, 6.15.28 immoral stimulus 1.2.1 importance sampling 4.11.12 inclusion-exclusion principle 1.3.4, 1.8.12 increasing sequence: of events 1.8.16; of r.v.s 2.7.2 increments: independent 9.7.6, 16--17; orthogonal 7.7.1; spectral 9.4.1, 3; stationary 9.7.17; of Wiener pr. 9.7.6 independence and symmetry 1.5.3 independent: conditionally 1.5 .5; continuous r.v.s 4.5.5, 4.14.6; current and excess life 10.6.9; customers 11.7 .1; discrete r. v.s 3 .11.1, 3; events 1.5 .1; increments 9.7.17; mean, variance of normal sample 4.10.5, 5.12.42; normal distn 4.7.5; pairwise 1.5.2, 3.2.1, 5.1.7; set 3.11.40; triplewise 5.1.7 indicators and matching 3.11.17 inequality: bivariate normal 4.7.12; Bonferroni 1.8.37; Boole 1.8.11; Cauchy-Schwarz 4.14.27; Chebyshov 7.11.9; Dobrushin 6.14.4; Doob-Kolmogorov 7.8.1; Doob L2 13.7.1; downcrossings 12.3.1; FKG 3.11.18; Hajek-Renyi-Chow 12.9.4-5; Hoeffding 12.2.1-2; Holder 4.14.27; Jensen 5.6.1, 7.9.4; Kolmogorov 7.8.1-2, 7.11.29-30; Kounias 1.8.38; Lyapunov 4.14.28; maximal 12.4.3-4, 12.9.3, 5, 9; m.g.f. 5.8.2, 12.9.7; Minkowski 4.14.27; triangle 7.1.1, 3; upcrossings 12.3.2 infinite divisibility 5.12.13-14 inner product 6.14.1, 7.1.2 inspection paradox 10.6.5 insurance 11.8.18, 12.9.12 integral: Monte Carlo 4.14.9; normal 4.14.1; stochastic 9.7.19, 13.8.1-2 invariant event 9.5.1 inverse square distn 3.1.1 inverse transform technique 2.3 .3 inversion theorem 5.9.5; c.f. 5.12.20
Index
5.12.33, 40, 7.11.26, 10.6.3; diffusion 13.3.1; distns 2.3.1; events 1.8.16; gamma 5.9.3; geometric-exponential 5.12.39; lim inf 1.8.16; lim sup 1.8.16, 5.6.3, 7.3.2, 9-10, 12; local J 5.9.2, 5.10.5-6; martingale Jaguar 3.11.25 7.8.3; normal 5.12.41, 7.11.19; Jensen's inequality 5.6.1, 7.9.4 Poisson 3.11.17; probability joint: c.f. 5.12.30; density 2.7.20; 1.8.16-18; r.v. 2.7.2; uniform distn 2.5.4; mass fn 2.5.5; 5.9.1 moments 5.12.30; p.g.f. 5.1.3-5 linear dependence 3.11.15 jump chain: of MIM!l 11.2.6 linear fn of normal r.v. 4.9.3-4 linear prediction 9. 7.1, 3 K local central limit theorem 5.9.2, key renewal theorem 10.3.3, 5, 5.10.5-6 10.6.11 logarithmic distn 3.1.1, 5.2.3 Keynes, J. M. 3.9.6 log-convex 3.1.5 knapsack problem 12.2.1 log-likelihood 7.11.31 Kolmogorov: criterion 6.5.2; log-normal r.v. 4.4.5, 5.12.43 inequality 7.8.1-2, 7.11.29-30 lottery 1.8.31 Korolyuk-Khinchin theorem 8.2.3 Lyapunov's inequality 4.14.28 Kounias's inequality 1.8.38 Krickeberg decomposition 12.9.11 M Kronecker's lemma 7.8.2, machine 11.8.17 7.11.30, 12.9.5 magnets 3.4.8 kurtosis 4.14.45 marginal: discontinuous 4.5.1; multinomial 3.6.2; order L statistics 4.14.22 L2 inequality 13.7.1 Markov chain in continuous Labouchere system 12.9.15 time: ergodic theorem 7 .11.33, lack of anticipation 6.9.4 10.5.1; first passage 6.9.5-6; lack-of-memory property: irreducible 6.15.15; jump chain exponential distn 4.14.5; 6.9.11; martingale 12.7.1; geometric distn 3 .11.7 mean first passage 6.9.6; mean recurrence time 6.9.11; ladders, see records renewal pr. 8.3.5; reversible Lancaster's theorem 4.14.38 6.15.16, 38; stationary distn large deviations 5.11.1-3, 12.9.7 6.9.11; two-state 6.9.1-2, last exits 6.2.1, 6.15.7 6.15.17; visits 6.9.9 lattice distn 5.7.5 Markov chain in discrete time: law: anomalous numbers 3.6.7; absorbing state 6.2.2; bivariate arc sine 3.10.3, 3.11.28, 5.3.5; 6.15.4; convergence 6.15.43; De Morgan 1.2.1; iterated dice 6.1.2; ergodic theorem logarithm 7.6.1; large numbers 7.11.32; finite 6.5.8, 6.15.43; 2.2.2; strong 7.4.1, 7.8.2, first passages 6.2.1, 6.3.6; 7.11.6, 9.7.10; unconscious homogeneous 6.1.1; imbedded statistician 3.11.3; weak 7.4.1, 6.9.11, 6.15.17, 11.4.1; last 7.11.15, 20-21; zero-one exits 6.2.1, 6.15.7; martingale 7.3.4-5 12.1.8, 12.3.3; mean first Lebesgue measure 6.15.29, passage 6.3.7; mean recurrence 13.12.16 time 6.9.11; persistent 6.4.10; left-continuous r.w. 5.3.7, 5.12.7 renewal 10.3.2; reversible level sets of Wiener pr. 13.12.16 6.14.1; sampled 6.1.4, 6.3.8; simulation of 6.14.3; stationary Levy metric 2.7.13, 7.1.4, 7.2.4 distn 6.9.11; sum 6.1.8; limit: binomial 3.11.1 0; two-state 6.15.11, 17, 8.2.1; binomial-Poisson 5.12.39; visits 6.2.3-5, 6.3.5, 6.15.5, 44 branching 12.9.8; central Markov-Kakutani theorem 6.6.1 limit theorem 5.10.1, 3, 9, irreducible Markov pr. 6.15.15 iterated logarithm 7 .6.1 Ito: formula 13.9.2; process 13.9.3
434
Markov process: Gaussian 9.6.2; reversible 6.15.16 Markov property 6.1.5, 10; strong 6.1.6 Markov renewal, see Poisson pr. Markov time, see stopping time Markovian queue, see M/M/1 marriage problem 3.4.3, 4.14.35 martingale: backward 12.7.3; birth-death 12.9.10; branching pr. 12.1.3, 9, 12.9.1-2, 8, 20; casino 7.7.4; continuous parameter 12.7.1-2; convergence 7.8.3, 12.1.5, 12.9.6; de Moivre 12.1.4, 12.4.6; exponential 13.3.9; finite stopping time 12.4.5; gambling 12.1.4, 12.5.8; Markov chain 12.1.8, 12.3.3, 12.7.1; optional stopping 12.5.1-8; orthogonal increments 7. 7.1; partial sum 12.7 .3; patterns 12.9.16; Poisson pr. 12.7.2; reversed 12.7.3; simple r.w. 12.1.4, 12.4.6, 12.5.4-7; stopping time 12.4.1, 5, 7; urn 7.11.27, 12.9.13-14; Wiener pr. 12.9.22-23 mass function, joint 2.5.5 matching 3.4.9, 3.11.17, 5.2.7, 12.9.21 matrix: covariance 3.11.15; definite 4.9.1; doubly stochastic 6.1.12, 6.15.2; multiplication 4.14.63; square root 4.9.1; stochastic 6.1.12, 6.14.1; sub-stochastic 6.1.12; transition 7.11.31; tridiagonal 6.5.1, 6.15.16 maximal: coupling 4.12.4-6, 7.11.16; inequality 12.4.3-4, 12.9.3, 5, 9 maximum of: branching pr. 12.9.20; multinormal 5.9.7; r.w. 3.10.2, 3.11.28, 5.3.1, 6.1.3, 12.4.6; uniforms 5.12.32; Wiener pr. 13.12.8, 11, 15, 17 maximum r.v. 4.2.2, 4.5.4, 4.14.17-18, 5.12.32, 7.11.14 mean: extreme-value 5.12.27; first passage 6.3.7, 6.9.6; negative binomial 5.12.4; normal 4.4.6; recurrence time 6.9.11; waiting time 11.4.2, 11.8.6, 10 measure: ergodic 9.7.11-12; Lebesgue 6.15.29, 13.12.16;
Index stationary 9.7.11-12; strongly mixing 9.7.12 median 2.7.11, 4.3.4, 7.3.11 menages 1.8.23 Mercator projection 6.13.5 meteorites 9. 7.4 metric 2.7.13, 7.1.4; Levy 2.7.13, 7 .1.4, 7 .2.4; total variation 2.7.13 m.g.f. inequality 5.8.2, 12.9.7 migration pr., open 11.7.1, 5 millionaires 3.9.4 Mills's ratio 4.4.8, 4.14.1 minimal, solution 6.3.6--7, 6.9.6 Minkowski's inequality 4.14.27 misprints 6.4.1 mixing, strong 9. 7.12 mixture 2.1.4, 2.3.4, 4.1.3, 5.1.9 modified renewall0.6.12 moments: branching pr. 5.4.1; fractional 3.3.5, 4.3.1; generating fn 5.1.8, 5.8.2, 5.11.3; joint 5.12.30; problem 5.12.43; renewal pr. 10.1.1; tail integral 4.3.3, 5.11.3 Monte Carlo 4.14.9 Monty Hall 1.4.5 Moscow 11.8.3 moving average 8.7.1, 9.1.3, 9.4.2, 9.5.3, 9.7.1-2, 7; spectral density 9. 7. 7 multinomial distn 3.5.1; marginals 3.6.2; p.g.f. 5.1.5 multinormal distn 4.9.2; c.f. 5.8.6; conditioned 4.9.6-7; covariance matrix 4.9.2; maximum 5.9.7; sampling from 4.14.62; standard 4.9.2; transformed 4.9.3, 4.14.62 Murphy's law 1.3.2 mutual information 3.6.5
N needle, Buffon's 4.5.2, 4.14.31-32 negative binomial distn 3.8.4; bivariate 5.12.16; moments 5.12.4; p.g.f. 5.1.1, 5.12.4 negative hypergeometric distn 3.5.4 Newton, I. 3.8.5 non-central distn 5.7.7-8 non-homogeneous: birth pr. 6.15.24; Poisson pr. 6.13.7, 6.15.19-20 noodle, Buffon's 4.14.31
norm 7.1.1, 7.9.6; equivalence class 7.1.1; mean square 7.9.6; rth mean 7.1.1 normal distn 4.4.6; bivariate 4.7.5-6, 12; central limit theory 5.10.1, 3, 9, 5.12.23, 40; characterization of 5.12.23; cumulants 5.7.4; limit 7.11.19; linear transfomations 4.9.3-4; Mills's ratio 4.4.8, 4.14.1; moments 4.4.6, 4.14.1; multivariate 4.9.2, 5.8.6; regression 4.8.7, 4.9.6, 4.14.13, 7.9.2; sample 4.10.5, 5.12.42; simulation of 4.11.7, 4.14.49; square 4.14.12; standard 4.7.5; sum 4.9.3; sum of squares 4.14.12; trivariate 4.9.8-9; uncorrelated 4.8.6 normal integral 4.14.1 normal number theorem 9.7.14 now 6.15.50
0 occupation time for Wiener pr. 13.12.20 open migration 11.7.1, 5 optimal: packing 12.2.1; price 12.9.19; reset time 13.12.22; serving 11.8.13 optimal stopping: dice 3.3.8-9; marriage 4.14.35 optional stopping 12.5.1-8, 12.9.19; diffusion 13.4.2; Poisson 12.7.2 order statistics 4.14.21; exponential 4.14.33; general 4.14.21; marginals 4.14.22; uniform 4.14.23-24, 39, 6.15.42, 12.7.3 Omstein-Uhlenbeck pr. 9.7.19, 13.3.4, 13.7.4-5, 13.12.3-4, 6; reflected 13.12.6 orthogonal: increments 7.7.1; polynomials 4.14.37 osmosis 6.15.36
p pairwise independent: events 1.5.2; r.v.s 3.2.1, 3.3.3, 5.1.7 paradox: Bertrand 4.14.8; Borel 4.6.1; Carroll 1.4.4; Galton 1.5.8; inspection 10.6.5; Parrando 6.15.48; St Petersburg 3.3.4; voter 3.6.6 parallel lines 4.14.52 parallelogram 4.14.60; property 7.1.2;
435
parking 4.14.30 Parrando's paradox 6.15.48 particle 6.15.33 partition of sample space 1.8.10 Pasta property 6.9.4 patterns 1.3.2, 5.2.6, 5.12.2, 10.6.17, 12.9.16 pawns 2.7.18 Pepys's problem 3.8.5 periodic state 6.5.4, 6.15.3 persistent: chain 6.4.10, 6.15.6; r.w. 5.12.5-6, 6.3.2, 6.9.8, 7.3.3, 13.11.2; state 6.2.3-4, 6.9.7 Petersburg, see St Petersburg pig 1.8.22 points, problem of 3.9.4, 3.11.24 Poisson: approximation 3.11.35; coupling 4.12.2; flips 3.5.2, 5.12.37; sampling 6.9.4 Poisson distn 3.5.3; characterization of 5.12.8, 15; compound 3.8.6, 5.12.13; and gamma distn 4.14.11; limit of binomial 5.12.39; modified 3.1.1; sum 3.11.6, 7.2.10 Poisson pr. 6.15.29; age 10.5.4; arrivals 8.7.4, 10.6.8; autocovariance 7 .11.5, 9 .6.1; characterization 6.15.29, 9.7.16; colouring theorem 6.15.39; compound 6.15.21; conditional property 6.13.6; continuity in m.s. 7 .11.5; covariance 7 .11.5; differentiability 7.11.5; doubly stochastic 6.15.22-23; excess life 6.8.3, 10.3.1; forest 6.15.30; gaps 10.1.2; Markov renewal pr. 8.3.5, 10.6.9; martingales 12.7.2; non-homogeneous 6.13.7, 6.15.19-20; optional stopping 12.7.2; perturbed 6.15.39-40; renewal 8.3.5, 10.6.9-10; Renyi's theorem 6.15.39; repairs 11.7.18; sampling 6.9.4; spatial 6.15.30-31, 7.4.3; spectral density 9.7.6; sphere 6.13.3-4; stationary increments 9.7.6, 16; superposed 6.8.1; thinned 6.8.2; total life 10.6.5; traffic 6.15.40, 49, 8.4.3 poker 1.8.33; dice 1.8.34 P6lya's urn 12.9.13-14 portfolio 13.12.23; self-financing 13.10.2-3 positive definite 9.6.1, 4.9.1
Index positive state, see non-null postage stamp lemma 6.3.9 potential theory 13.11.1-3 power series approximation 7.11.17 Pratt's lemma 7.10.5 predictable step fn 13.8.4 predictor: best 7.9.1; linear 7.9.3, 9.2.1-2, 9.7.1, 3 probabilistic method 1.8.28, 3.4. 7 probability: continuity 1.8.16, 1.8.18; p.g.f. 5.12.4, 13; vector 4.11.6 problem: matching 3.4.9, 3.11.17, 5.2.7, 12.9.21; menages 1.8.23; Pepys 3.8.5; of points 3.9.4, 3.11.24; Waldegrave 5.12.10 program, dual, linear, and primal 6.6.3 projected r.w. 5.12.6 projection theorem 9.2.10 proof-reading 6.4.1 proportion, see empirical ratio proportional investor 13.12.23 prosecutor's fallacy 1.4.6 protocol 1.4.5, 1.8.26 pull-through property 3.7.1
Q quadratic variation 8.5.4, 13.7.2 queue: batch 11.8.4; baulking 8.4.4, 11.8.2, 19; busy period 6.12.1, 11.3.2-3, 11.5.1, 11.8.5, 9; costs 11.8.13; departure pr. 11.2.7, 11.7.2-4, 11.8.12; difficult customer 11.7.4; D/M/1 11.4.3, 11.8.15; dual 11.5.2; Erlang's loss fn 11.8.19; finite waiting room 11.8.1; G/G/1 11.5.1, 11.8.8; G/M/1 11.4.1-2, 11.5.2-3; heavy traffic 11.6.1, 11.8.15; idle period 11.5.2, 11.8.9; imbedded branching 11.3.2, 11.8.5, 11; imbedded Markov pr. 11.2.6, 11.4.1, 3, 11.4.1; imbedded renewal 11.3.3, 11.5.1; imbedded r.w. 11.2.2, 5; Markov, see M/M/1; M/D/1 11.3.1, 11.8.10-11; M/G/1 11.3.3, 11.8.6-7; M/G/oo 6.12.4, 11.8.9; migration system 11.7.1, 5; M/M/1 6.9.3, 6.12.1, 11.2.2-3, 5-6, 11.3.2, 11.6.1, 11.8.5, 12; M!Mik 11.7.2, 11.8.13; M/M/oo 8.7.4; series 11.8.3, 12; supermarket 11.8.3; tandem 11.2.7, 11.8.3; taxicabs 11.8.16; telephone
orthogonal 7.7.1; p.g.f. 5.12.4, 13; Poisson 3.5.3; standard normal 4.7.5; Student's t 4.10.2-3, 5.7.8; symmetric 3.2.5, 4.1.2, 5.12.22; tails 3.11.13, 4.3.3, 5, 4.14.3, 5.1.2, 5.6.4, 5.11.3; tilted 5.1.9, 5.7.11; trivial3.11.2; truncated 2.4.2; uncorrelated 3.11.12, 16, 4.5.7-8, 4.8.6; uniform 3.8.1, R 4.8.4, 4.11.1, 9.7.5; waiting radioactivity 10.6.6-8 time, see queue; Weibull 4.4.7; zeta or Zipf 3.11.5 random: bias 5.10.9; binomial coefficient 5 .2.1; chord random walk: absorbed 3.11.39, 4.13.1; dead period 10.6.7; 12.5.4-5; arc sine laws harmonic series 7.11.37; 3.10.3, 3.11.28, 5.3.5; on integers 6.15.34; line 4.13.1-3, binary tree 6.4.7; conditional 4.14.52; paper 4.14.56; 3.9.2-3; on cube 6.3.4; parameter 4.6.5, 5.1.6, first passage 5.3.8; first 5.2.3, 5.2.8; particles 6.4.8; visit 3.10.3; on graph 6.4.6, pebbles 4.14.51; permutation 9, 13.11.2-3; on hexagon 4.11.2; perpendicular 4.14.50; 6.15.35; imbedded in queue polygons 4.13.10, 6.4.9; rock 11.2.2, 5; left-continuous 5.3.7, 4.14.57; rods 4.14.25-26, 5.12.7; martingale 12.1.4, 53-54; sample 4.14.21; 12.4.6, 12.5.4-5; maximum subsequence 7.11.25; sum 3.10.2, 3.11.28, 5.3.1; 3.7.6, 3.8.6, 5.2.3, 5.12.50, persistent 5.12.5-6, 6.3.2; 10.2.2; telegraph 8.2.2; triangle potentials 13.11.2-3; projected 4.5.6, 4.13.6-8, 11, 13; 5.12.6; range of 3.11.27; velocity 6.15.33, 40 reflected 11.2.1-2; retaining barrier 11.2.4; returns to origin random sample: normal 4.10.5; 3.10.1, 5.3.2; reversible 6.5.1; ordered 4.12.21 simple 3.9.1-3, 5, 3.10.1-3; on random variable: see also density square 5.3.3; symmetric 1.7.3, and distribution; arc sine 3.11.23; three dimensional 4.11.13; arithmetic 5.9.4; 6.15.9-10; transient 5.12.44, Bernoulli 3.11.14, 35; beta 6.15.9, 7.5.3; truncated 6.5.7; 4.11.4; beta-binomial 4.6.5; two dimensional 5.3.4, 5.12.6, binomial 2.1.3, 3.11.8, 11, 12.9.17; visits 3.11.23, 6.9.8, 5.12.39; bivariate normal 10; zero mean 7.5.3; zeros of 4.7.5-6, 12; Cauchy 4.4.4; 3.10.1, 5.3.2, 5.12.5-6 c.f. 5.12.26-31; chi-squared range of r.w. 3.11.27 4.10.1; compounding rate of convergence 6.15.43 5.2.3; continuous 2.3.1; ratios 4.3.2; Mills's 4.4.8; sex Dirichlet 3.11.31, 4.14.58; expectation 5.6.2, 7.2.3; 3.11.22 exponential 4.4.3, 5.12.39; record: times 4.2.1, 4, 4.6.6, 10; extreme-value 4.1.1, 4.14.46; values 6.15.20, 7.11.36 F(r, s) 4.10.2, 4, 5.7.8; recurrence, see difference gamma 4.11.3, 4.14.10-12; recurrence time 6.9 .11 geometric 3.1.1, 3.11.7; recurrent: event 5.12.45, 7.5.2, hypergeometric 3 .11.1 0-11; 9.7.4; see persistent independent 3.11.1, 3, 4.5.5, red now 12.9.18 7.2.3; indicator 3.11.17; infinitely divisible 5.12.13-14; reflecting barrier: s.r.w. 11.2.1-2; drifting Wiener pr. 13.5.1; logarithmic 3.1.1, 5.2.3; Ornstein-Uhlenbeck pr. log-normal 4.4.5; median 13.12.6 2.7.11, 4.3.4; m.g.f. 5.1.8, regeneration 11.3 .3 5.8.2, 5.11.3; multinomial regression 4.8.7, 4.9.6, 4.14.13, 3.5.1; multinormal 4.9.2; 7.9.2 negative binomial 3.8.4; rejection method 4.11.3-4, 13 normal 4.4.6, 4.7.5-6, 12; exchange 11.8.9; two servers 8.4.1, 5, 11.7.3, 11.8.14; virtual waiting 11.8.7; waiting time 11.2.3, 11.5.2-3, 11.8.6, 8, 10 quotient 3.3.1, 4.7.2, 10, 13-14, 4.10.4, 4.11.10, 4.14.11, 14, 16, 40, 5.2.4, 5.12.49, 6.15.42
436
Index reliability 3.4.5-6, 3.11.18-20, renewal: age, see current life; alternating 10.5.2, 10.6.14; asymptotics 10.6.11; Bernoulli 8.7.3; central limit theorem 10.6.3; counters 10.6.6-8, 15; current life 10.3.2, 10.5.4, 10.6.4; delayed 10.6.12; excess life 8.3.2, 10.3.1-4, 10.5.4; r. function 10.6.11; gaps 10.1.2; key r. theorem 10.3.3, 5, 10.6.11; Markov 8.3.5; rn.g.f. 10.1.1; moments 10.1.1; Poisson 8.3.5, 10.6.9-10; r. process 8.3.4; r.-reward 10.5.1-4; r. sequence 6.15.8, 8.3.1, 3; stationary 10.6.18; stopping time 12.4.2; sum/superposed 10.6.10; thinning 10.6.16 Renyi's theorem 6.15.39 repairman 11.7.18 repulsion 1.8.29 reservoir 6.4.3 resources 6.15.47 retaining barrier 11.2.4 reversible: birth-death pr. 6.15.16; chain 6.14.1; Markov pr. 6.15.16, 38; queue 11.7.2-3, 11.8.12, 14; r.w. 6.5.1 Riemarm-Lebesgue lemma 5.7.6 robots 3.7.7 rods 4.14.25-26, ruin 11.8.18, 12.9.12; see also gambler's ruin runs 1.8.21, 3.4.1, 3.7.10, 5.12.3, 46-47
s a-field 1.2.2, 4, 1.8.3, 9.5.1, 9.7.13; increasing sequence of 12.4.7 StJohn's College 4.14.51 St Petersburg paradox 3.3.4 sample: normal 4.10.5, 5.12.42; ordered 4.12.21 sampling 3.11.36; Poisson 6.9.4; with and without replacement 3.11.10 sampling from distn: arc sine 4.11.13; beta 4.11.4-5; Cauchy 4.11.9; exponential4.14.48; gamma 4.11.3; geometric 4.11.8; Markov chain 6.14.3; multinormal 4.14.62; normal 4.11.7, 4.14.49; s.r.w. 4.11.6; uniform 4 .11.1
secretary problem 3.11.17, 4.14.35 self-financing portfolio 13.10.2-3 semi-invariant 5. 7.3-4 sequence: of c.f.s 5.12.35; of distns 2.3.1; of events 1.8.16; of heads and tails, see pattern; renewal 6.15.8, 8.3.1, 3; of r.v.s 2.7.2 series of queues 11.8.3, 12 shift operator 9.7.11-12 shocks 6.13.6 shorting 13.11.2 simple birth pr. 6.8.4-5, 6.15.23 simple birth-death pr.: conditioned 6.11.4-5; diffusion approximation 13.3.1; extinction 6.11.3, 6.15.27; visits 6.11.6-7 simple immigration-death pr. 6.11.2, 6.15.18 simple: process 8.2.3; r.w. 3.9.1-3, 5, 3.10.1-3, 3.11.23, 27-29, 11.2.1, 12.1.4, 12.4.6, 12.5.4-7 simplex 6.15.42; algorithm 3.11.33 simulation, see sampling sixes 3.2.4 skewness 4.14.44 sleuth 3.11.21 Slutsky's theorem 7.2.5 smoothing 9.7.2 snow 1.7.1 space, vector 2.7.3, 3.6.1 span of r.v. 5.7.5, 5.9.4 Sparre Andersen theorem 13.12.18 spectral: density 9.3 .3; distribution 9.3.2, 4, 9.7.2-7; increments 9.4.1, 3 spectrum 9.3.1 sphere 1.8.28, 4.6.1, 6.13.3-4, 12.9.23, 13.11.1; empty 6.15.31 squeezing 4.14.47 standard: bivariate normal 4.7.5; multinormal 4.9.2; normal 4.7.5; Wiener pr. 9.6.1, 9.7.18-21, 13.12.1-3 state: absorbing 6.2.2; persistent 6.2.3-4; symmetric 6.2.5; transient 6.2.4 stationary distn 6.9.1, 3-4, 11-12; 6.11.2; birth-death pr. 6.11.4; current life 10.6.4; excess life 10.3.3; Markov chain 9.1.4; open migration 11.7.1,
437
5; queue length 8.4.4, 8.7.4, 11.2.1-2, 6, 11.4.1, 11.5.2, and Section 11.8 passim; r.w. 11.2.1-2; waiting time 11.2.3, 11.5.2-3, 11.8.8 stationary excess life 10.3.3 stationary increments 9. 7.1 7 stationary measure 9.7.11-12 stationary renewal pr. 10.6.18, Stirling's formula 3.10.1, 3.11.22, 5.9.6, 5.12.5, 6.15.9, 7.11.26 stochastic: integral 9.7.19, 13.8.1-2; matrix 6.1.12, 6.14.1, 6.15.2; ordering 4.12.1-2 stopping time 6.1.6, 10.2.2, 12.4.1-2, 5, 7; for renewal pr. 12.4.2 strategy 3.3.8-9, 3.11.25, 4.14.35, 6.15.50, 12.9.19, 13.12.22 strong law of large numbers 7.4.1, 7.5.1-3, 7.8.2, 7.11.6, 9.7.10 strong Markov property 6.1.6 strong mixing 9.7.12 Student's t distn 4.10.2-3; non-central 5.7.8 subadditive fn 6.15.14, 8.3.3 subgraph 13.11.3 sum of independent r.v.s: Bernoulli 3.11.14, 35; binomial 3.11.8, 11; Cauchy 4.8.2, 5.11.4, 5.12.24-25; chi-squared 4.10.1, 4.14.12; exponential 4.8.1, 4, 4.14.10, 5.12.50, 6.15.42; gamma 4.14.11; geometric 3.8.3-4; normal 4.9.3; p.g.f. 5.12.1; Poisson 3.11.6, 7.2.10; random 3.7.6, 3.8.6, 5.2.3, 5.12.50, 10.2.2; renewals 10.6.10; uniform 3.8.1, 4.8.5 sum of Markov chains 6.1.8 supercritical branching 6. 7.2 supermartingale 12.1.8 superposed: Poisson pr. 6.8.1; renewal pr. 10.6.10 sure thing principle 1. 7.4 survival 3.4.3, 4.1.4 Sylvester's problem 4.13.12, 4.14.60 symmetric: r.v. 3.2.5, 4.1.2, 5.12.22; r.w. 1.7.3; state 6.2.5 symmetry and independence 1.5.3 system 7.7.4; Labouchere 12.9.15
Index
T
two-dimensional: r.w. 5.3.4, 5.12.6, 12.9.17; Wiener pr. 13.12.12-14 two server queue 8.4.1, 5, 11.7.3, 11.8.14 two-state Markov chain 6.15.11, 17, 8.2.1; Markov pr. 6.9.1-2, 6.15.16-17 Type: T. one counter 10.6.6-7; T. two counter 10.6.15
11.4.2; in M/D/1 11.8.1 0; in M/G/1 11.8.6; in MIM/1 11.2.3; stationary distn 11.2.3, 5.7.8 11.5.2-3; virtual 11.8.7 tail: c.f. 5.7.6; equivalent 7.11.34; Wald's eqn 10.2.2-3 event 7.3.3, 5; function 9.5.3; integral 4.3.3, 5; sum 3.11.13, Waldegrave's problem 5.12.10 4.14.3 Waring's theorem 1.8.13, 5.2.1 tail of distn: and moments 5.6.4, weak law of large numbers 7.4.1, 5.11.3; p.g.f. 5.1.2 7.11.15, 20-21 tandem queue 11.2.7, 11.8.3 Weibull distn 4.4.7, 7.11.13 taxis 11.8.16 u Weierstrass's theorem 7.3.6 telekinesis 2. 7.8 u Mysaka 8.7.7 white noise, Gaussian 13.8.5 telephone: exchange 11.8.9; sales unconscious statistician 3.11.3 3.11.38 Wiener process: absorbing uncorrelated r.v. 3.11.12, 16, barriers 13.13.8, 9; arc testimony 1.8.27 4.5.7-8, 4.8.6 sine laws 13.4.3, 13.12.10, thinning 6.8.2; renewal 10.6.16 uniform integrability: Section 13.12.19; area 12.9.22; three-dimensional r.w. 6.15.10; 7.10 passim, 10.2.4, 12.5.1-2 Bartlett eqn 13.3.3; Bessel transience of 6.15.9, uniform distn 3.8.1, 4.8.4, 4.11.1, pr. 13.3.5; on circle 13.9.4; three-dimensional Wiener pr. 9.7.5; maximum 5.12.32; conditional 8.5.2, 9.7.21, 13.11.1 order statistics 4.14.23-24, 39, 13.6.1; constructed 13.12.7; 6.15.42; sample from 4.11.1; three series theorem 7.11.35 d-dimensional 13.7.1; drift sum 3.8.1, 4.8.4 13.3.3, 13.5.1; on ellipse tied-down Wiener pr. 9.7.21-22, 13.9.5; expansion 13.12.7; uniqueness of conditional 13.6.2-5 first passage 13.4.2; geometric expectation 3.7.2 tilted distn 5.1.9, 5.7.11 13.3.9, 13.4.1; hits sphere time-reversibility 6.5.1-3, 6.15 ..16 upcrossings inequality 12.3.2 13.11.1; hitting barrier upper class fn 7.6.1 total life 10.6.5 13.4.2; integrated 9.7.20, urns 1.4.4, 1.8.24-25, 3.4.2, total variation distance 4.12.3-4, 12.9.22, 13.3.8, 13.8.1-2; level 4, 6.3.10, 6.15.12; P6lya's 7.2.9, 7.11.16 sets 13.12.16; martingales 12.9.13-14 12.9.22-23, 13.3.8-9; tower property 3.7.1, 4.14.29 maximum 13.12.8, 11, 15, traffic: gaps 5.12.45, 8.4.3; 17; occupation time 13.12.20; heavy 11.6.1, 11.8.15; Poisson value function 13.10.5 quadratic variation 8.5.4, 6.15.40, 49, 8.4.3 variance: branching pr. 5.12.9; 13.7.2; reflected 13.5.1; sign transform, inverse 2.3.3 conditional 3.7.4, 4.6.7; normal 13.12.21; standard 9.7.18-21; transient: r.w. 5.12.44, 7.5.3; 4.4.6 three-dimensional 13.11.1; Wiener pr. 13.11.1 tied-down, see Brownian vector space 2.7.3, 3.6.1 transition matrix 7.11.31 bridge; transformed 9.7.18, Vice-Chancellor 1.3.4, 1.8.13 13.12.1, 3; two-dimensional transitive coins 2.7.16 virtual waiting 11.8.7 13.12.12-14; zeros of 13.4.3, trapezoidal distn 3.8.1 visits: birth-death pr. 6.11.6-7; 13.12.10 Markov chain 6.2.3-5, 6.3.5, trial, de Moivre 3.5.1 Wiener-Hopf eqn 11.5.3, 11.8.8 r.w. 3.11.23, 6.9.9, 6.15.5, 44; triangle inequality 7 .1.1, 3 29, 6.9.8, 10 Trinity College 12.9.15 voter paradox 3.6.6 X triplewise independent 5.1. 7 X-ray 4.14.32 trivariate normal distn 4.9.8-9 trivial r. v. 3.11.2 waiting room 11.8.1, 15 truncated: r.v. 2.4.2; r.w. 6.5.7 waiting time: dependent 11.2.7; Tunin's theorem 3.11.40 zero-one law, Hewitt-Savage for a gap 10.1.2; in G/G/1 turning time 4.6.10 11.5.2, 11.8.8; in G/M/1 7.3.4-5 t, Student's 4.10.2-3; non-central
v
w
z
438
FROM A REVIEW OF PROBABILITY AND RANDOM PROCESSES: PROBLEMS AND SOLUTIONS