MARKOV PROCESSES: THEORY AND EXAMPLES
JAN SWART AND ANITA WINTER
Date: April 10, 2013. 1 2 JAN SWART AND ANITA WINTER
Contents 1. Stochastic processes 3 1.1. Random variables 3 1.2. Stochastic processes 5 1.3. Cadlag sample paths 6 1.4. Compactification of Polish spaces 18 2. Markov processes 23 2.1. The Markov property 23 2.2. Transition probabilities 27 2.3. Transition functions and Markov semigroups 30 2.4. Forward and backward equations 32 3. Feller semigroups 34 3.1. Weak convergence 34 3.2. Continuous kernels and Feller semigroups 35 3.3. Banach space calculus 37 3.4. Semigroups and generators 40 3.5. Dissipativity and the maximum principle 42 3.6. Hille-Yosida: different formulations 46 3.7. Dissipative operators 48 3.8. Resolvents 50 3.9. Hille-Yosida: proofs 51 4. Feller processes 56 4.1. Markov processes 56 4.2. Jump processes 57 4.3. Fellerprocesseswithcompactstatespace 62 4.4. Feller processes with locally compact state space 65 5. Harmonic functions and martingales 70 5.1. Harmonic functions 70 5.2. Filtrations 70 5.3. Martingales 72 5.4. Stopping times 74 5.5. Applications 76 5.6. Non-explosion 79 6. Convergence of Markov processes 81 6.1. Convergence in path space 81 6.2. Proof of the main result (Theorem 4.2) 87 7. Strong Markov property 89 References 93 MARKOV PROCESSES 3
1. Stochastic processes In this section we recall some basic definitions and facts on topologies and stochastic processes (Subsections 1.1 and 1.2). Subsection 1.3 is devoted to the study of the space of paths which are continuous from the right and have limits from the left. Finally, for sake of completeness, we collect facts on compactifications in Subsection 1.4. These will only find applications in later sections. 1.1. Random variables. Probability theory is the theory of random vari- ables, i.e., quantities whose value is determined by chance. Mathemati- cally speaking, a random variable is a measurable map X :Ω E, where (Ω, , P) is a probability space and (E, ) is a measurable space.→ The prob- abilityF measure E 1 P = (X) := P X− X L ◦ on (E, ) is called the law of X and usually the only object that we are really E 1 interested in. If (Xt)t T is a family of random variables, taking values in ∈ measurable spaces (Et, t)t T , then we can view (Xt)t T as a single random E ∈ ∈ variable, taking values in the product space t T Et equipped with the 2 ∈ product-σ-field t T t. The law P(Xt)t∈T = ((Xt)t T ) of this random ∈ E QL ∈ variable is called the joint law of the random variables (Xt)t T . In practise, weQ usually need a bit more structure on the space∈ s that our random variables take values in. For our purposes, it will be sufficient to consider random variables taking values in Polish spaces. Recall that a topology on a space E is a collection of subsets of E, called open sets, such that: O (1) E, . ∅ ∈ O (2) Ot t T implies t T Ot . ∈O∀ ∈ ∈ ∈ O (3) O1, O2 implies O1 O2 . ∈ O ∩S ∈ O A topology is metrizable if there exists a metric d on E such that the open sets in this topology are the sets O with the property that x O ε > ∀ ∈ ∃ 0 s.t. Bε(x) O, where Bε(x) := y E : d(x,y) < ε is the open ball around x with⊂ radius ε. Two metrics{ are∈ called equivalent}if they define the same topology. Concepts such as convergence, continuity, and compactness depend only on the topology but completeness depends on the choice of the
1At this point, one may wonder why probabilists speak of a random variables at all and do not immediately focus on the probability measures that are their laws, if that is what they are really after. The reason is mainly a matter of convenient notation. If µ = L(X) is the law of a real-valued random variable X, then what is the law of X2? In terms of random variables, this is simply L(X2). In terms of probability measures, this is the image of the probability measure µ under the map x 7→ x2, i.e., the measure µ ◦ f −1 where f : R → R is defined as f(x)= x2 –an unpleasantly long mouthful. 2 Recall that Qt∈T Et := {(xt)t∈T : xt ∈ Et ∀t ∈ T }. The coordinate projections πt : Qt∈T Et := {(xt)t∈T : xt ∈ Et} → Et are defined by πt((xs)s∈T ) := xt, t ∈ T . By definition, the product-σ-field Qt∈T Et is the σ-field on Qt∈T Et that is generated by the −1 coordinate projections, i.e., Qt∈T Et := σ(πt : t ∈ T )= σ({πt (A) : A ∈Et}). 4 JAN SWART AND ANITA WINTER metric.3 A topological space E is called separable if there exists a countable set D E such that D is dense in E. By⊂ definition, a topological space (E, ) is Polish if E is separable and there exists a complete metric defining theO topology on E. We always equip Polish spaces with the Borel-σ-field (E), which is the σ-field generated by the open sets. B The reason why we are interested in Polish spaces is that for random variables taking values in Polish spaces, certain useful results are true that do not hold in general, since they make use of the fact. Lemma 1.1 (Probability measures on Polish spaces are tight). Each prob- ability measure P on a Polish space (E, ) is tight, i.e., for all ε> 0 there is a compact set K E such that P(K)O 1 ε. ⊆ ≥ −
Proof. Let (xk)k N be dense in (E, ), and let P be a probability measure ∈ O on (E, ). Given ε> 0 and a metric d on (E, ), we can choose N1,N2,... such thatO O
Nn 1 ε (1.1) P x′ : d(x′,x ) < 1 . ∪k=1 { k n} ≥ − 2n Nn 1 Let K be the closure of n 1 k=1 x′ : d(x′,xk) < n . Then K is totally 4 ≥ { } n bounded , and hence compact, and we have P(K) 1 ε ∞ 2 = 1 ε. T S ≥ − n=1 − − P For example, the following result states that provided the state space (E, ) is Polish, for each projective family of probability measures there existsO a projective limit. Theorem 1.2 (Percy J. Daniell [Dan19], Andrei N. Kolmogorov [Kol33]). Let (Et)t T be (a possibly uncountable) collection of Polish spaces and let ∈ µS (S T finite) be probability measures on (Et)t S such that ⊂ ∈ 1 (1.2) µ ′ (π )− = µ , S S′ T, S,S′ finite, S ◦ S S ⊂ ⊂ where πS denotes the projection on (Et)t S . Then there exists a unique ∈ probability measure µT on t T Et, equipped with the product σ-field, such that ∈ Q 1 (1.3) µ π− = µ , S T, S finite. T ◦ S S ⊂
Proof. For E R see e.g. Theorem 2.2.2 in [KS88]. t ≡
3More precisely: completeness depends on the uniform structure defined by the metric. For the theory of uniform spaces, see for example [Kel55]. 4Recall that a set A is totally bounded if for each ε > 0, A possesses a finite ε-net, where an ε-net for A is a collection of points {xn} with the property that for each x ∈ A there is an xk such that d(x,xk) <ε. MARKOV PROCESSES 5
A consequence of Kolmogorov’s extension theorem is that if µS : S T finite are probability measures satisfying the consistency relat{ion (1.2),⊂ } then there exist random variables (Xt)t T defined on some probability space ∈ (Ω, , P) such that ((Xt)t S )= µS for each finite S T . (The canonical F L ∈ ⊂ choice is Ω = t T Et.) ∈ Exercise 1.3.QFor n N, ki 0, 1 , i = 0,...,n, and 0 =: t0 < t1 < tτ −1 tτ (1.4) e− n e− n , if τn n, := 1 0 k1 kn 1 −tn ≤ { ≤ ≤···≤ ≤ } e− , if τn =1+ n. t1,...,tn (i) Show that the collection µ ; 0 =: t0 < t1 < < tn of prob- { n · · · } ability measures on k 0, 1 : 0 k1 kn 1 satisfies the consistence condition{ ∈ (1.2).{ } ≤ ≤···≤ ≤ } (ii) Can you find one (or even more than one) 0, 1 -valued stochastic process X with { } (1.5) P X = k ,...,X = k = µt1,...,tn (k ,...,k ), 0 k k 1. { t1 1 tn n} 1 n ≤ 1 ≤···≤ n ≤ 1.2. Stochastic processes. A stochastic process with index set T and state space E is a collection of random variables X = (Xt)t T (defined on a probability space (Ω, , P)) with values in E. We will usually∈ be interested in the case that T =F [0, ) and E is a Polish space. We interpret X = ∞ (Xt)t [0, ) as a quantity the value of which is determined by chance and that develops∈ ∞ in time. A stochastic process is called measurable if the map (t,ω) X (ω) from 7→ t [0, ) Ω into E is measurable. The functions t Xt(ω) (with ω Ω) are called∞ × the sample paths of the process X. 7→ ∈ Lemma 1.4 (Right continuous sample paths). If X has right continuous sample paths then X is measurable. (n) (n) Proof. Define processes X by Xt := X nt+1 /n. Then, for each measur- ⌊ ⌋ (n) 1 able set A E, (t,ω) : X (ω) A = ∞ [k/n, (k + 1)/n) X− (A), ⊆ { t ∈ } k=0 × k/n (n) so X is measurable for each n 1. By theS right-continuity of the sample paths, X(n) X pointwise, so≥X is measurable. n −→→∞ By definition, the laws (Xt1 ,...,Xtn ) with 0 t1 < < tn are called the finite-dimensional distributionsL of X. If X and≤Y are stochastic· · · processes with the same finite dimensional distributions then we say that Y is a version of X (and vice versa). Here X and Y need not be defined on the same probability space. If X and Y are stochastic processes defined on the same probability space then we say that Y is a modification of X if Xt = Yt a.s. 6 JAN SWART AND ANITA WINTER t 0. Note that if Y is a modification of X, then X and Y have the same finite∀ ≥ dimensional distributions. We say that X and Y are indistinguishable if X = Y t 0 a.s.5 t t ∀ ≥ Example (Modification). Let (Ω, , P) = ([0, 1], [0, 1],ℓ) where ℓ is the Lebesgue measure. For a given xF (0, ), defineB [0, 1]-valued stochastic processes Xx and Y by ∈ ∞ t, if t = x, (1.6) Xx(t) := 0, if t 6= x, and (1.7) Y (t) := t, t [0, 1]. ∈ Then Y is a modification of Xx but X and Y are not indistinguishable. Lemma 1.5 (Right continuous modifications). If Y is a modification of X and X and Y have right-continuous sample paths, then X and Y are indistinguishable. Proof. If Y is a modification of X then Xt = Yt t Q a.s. By right- continuity of sample paths, this implies that X = Y∀ t∈ 0 a.s. t t ∀ ≥ We will usually be interested in stochastic processes with sample paths that have right limits Xt+ := lims t Xs for each t 0 and left limits ↓ ≥ Xt := lims t Xs for each t > 0. In practise nobody can measure time with− infinite↑ precision, so when we model a real process it is a matter of taste whether we assume that the sample paths are right or left continuous; it is tradition to assume that they are right continuous. (Lemmas 1.4 and 1.5 hold equally well for processes with left continuous sample paths.) Note that a consequence of this assumption is that the sample paths cannot have a jump at time t = 0; this will actually be convenient later on. In the next section we study the space of all paths that are right-continuous with left limits in more detail. 1.3. Cadlag sample paths. Let (E, ) be a metrizable space. A function w : [0, ) E such that w is right continuousO and w(t ) exists for each t> 0 is∞ called→ a cadlag function (from the French “continu− `a droit limite `a gauche”). The space of all such functions is denoted by E[0, ) := (1.8) D ∞ w : [0, ) E : w(t)= w(t+) t 0, w(t ) exists t> 0 . ∞ → ∀ ≥ − ∀ 5Note the order of the statements: If Y is a modification of X, then there is for each ∗ P ∗ ∗ t ≥ 0 a measurable set Ωt ⊂ Ω with (Ω ) = 1 such that Xt(ω)= Yt(ω) for all ω ∈ Ωt . If X and Y are indistiguishable, then there exists a measurable set Ω∗ (independent of t) ∗ such that Xt(ω)= Yt(ω) for all ω ∈ Ω . MARKOV PROCESSES 7 We begin by observing that functions in E[0, ) are better behaved than one might suspect. D ∞ Lemma 1.6 (Only countably many jumps). If w E[0, ), then w has at most countably many points of discontinuity. ∈ D ∞ Proof. For n = 1, 2,..., and d a metric on (E, ), let O 1 (1.9) A := t> 0 : d w(t), w(t ) > . n − n Since w has limits from the right and the left, An can not possess clus- ter points. Hence An is countable for all n = 1, 2,..., and the set of all discontinuities of n 1An is countable too. ∪ ≥ In order to be in a position to do probability on spaces of random variables with values in [0, ) we want to equip [0, ) with a topology so that DE ∞ DE ∞ in this topology E[0, ) is Polish. We wil see that this is possible provided that (E, ) is Polish.D ∞ To motivateO the topology that we will choose, we first take a look at the space (1.10) [0, ) := continuous functions w : [0, ) E . CE ∞ ∞ → Lemma 1.7 (Uniform convergence on compacta). Let (E, d) be metric space. Then the following conditions on functions wn, w E[0, ) are equivalent. ∈ C ∞ (a) For all T > 0, (1.11) lim sup d(wn(t), w(t)) = 0. n t [0,T ] →∞ ∈ (b) For all (tn)n N,t [0, ) such that tn t, ∈ ∈ ∞ n−→→∞ (1.12) lim wn(tn)= w(t). n →∞ Proof. (a) (b). If tn t then there is a T > 0 such that tn,t T for all n. Now ⇒ → ≤ d w (t ), w(t) d w(t ), w(t) + d w (t ), w(t ) n n ≤ n n n n (1.13) d w(tn), w(t) + sup d wn(s), w(s) 0. ≤ s [0,T ] n−→→∞ ∈ by (a) and the continuity of w. (b) (a). Imagine that there exists a T > 0 such that ⇒ (1.14) lim sup sup d(wn(t), w(t)) = ε> 0. n t [0,T ] →∞ ∈ Then we can choose sn [0, T ] such that lim supn d(wn(sn), w(sn)) = ∈ →∞ ε. By the compactness of [0, T ] we can choose n1 < n2 < such that · · · ε limm snm = t for some t [0, T ] and d(wnm (snm ), w(snm )) for each →∞ ∈ ≥ 2 8 JAN SWART AND ANITA WINTER m. Hence, d wnm (snm ), w(t) + d w(snm ), w(t) d(wnm (snm ), w(snm )) ε ≥ ≥ 2 . By continuity, d(w(snm ), w(t)) 0. We therefore find that m −→→∞ ε (1.15) lim sup d(wn(sn), w(t)) n ≥ 2 →∞ which contradicts (1.12). If wn, w are as in Lemma 1.7 then because of Property (a) we say that wn converges to w uniformly on compacta. Property (b) shows that this definition does not depend on the choice of the metric on E, i.e., if d and d˜are equivalent metrics on E then w w uniformly on compacta w.r.t. d if and n → only if wn w uniformly on compacta w.r.t. d˜. The topology on E[0, ) of uniform→ convergence on compacta is metrizable. A possible choiceC ∞ of a metric on E[0, ) generating the topology of uniform convergence on compacta is forC example:∞ ∞ s (1.16) d (w , w ) := ds e− sup 1 d w (t s), w (t s) u.c. 1 2 ∧ 1 ∧ 2 ∧ Z0 t [0, ) ∈ ∞ Remark. If d is a metric on (E, ), then 1 d is also a metric, and both metrics are equivalent. O ∧ On E[0, ), we could also define uniform convergence on compacta as in LemmaD 1.7,∞ Property (a), but this topology would be too strong for our purposes. For example, if E = R, we would like the functions wn := 1[1+ 1 , ) n ∞ to approximate the function w := 1[1, ) as n , but supt [0,2] wn(t) ∞ → ∞ ∈ | − w(t) = 1 for each n. We wish to find a topology on [0, ) such that | DE ∞ wn w whenever the jump times of the functions wn converge to the jump times→ of w while the “rest” of the paths converge uniformly in compacta. The main result of this section is that such a topology exists and has nice properties. Theorem 1.8 (Skorohod topology). Let (E, d) be a metric space. Then there exists a metric dd on [0, ) such that in this metric, [0, ) Sk DE ∞ DE ∞ is separable if E is separable, E[0, ) is complete if E is complete, and w w if and only if for all T D [0, ∞) there exists a sequence λ of strictly n → ∈ ∞ n increasing, continuous functions λn : [0, T ] [0, ) with λn(0) = 0, such that → ∞ (1.17) lim sup λn(t) t = 0 n t [0,T ] | − | →∞ ∈ and for (tn)n N,t [0, T ], ∈ ∈ w(t), whenever (tn) t, (1.18) lim wn(λn(tn)) = ↓ n w(t ), whenever (tn) t. →∞ − ↑ MARKOV PROCESSES 9 Remark. The idea of the functions λn in (1.17) and (1.18) is to make two functions w, w˜ close in the topology on E[0, ) if a small deformation of the time scale makes them near in the uniformD ∞ topology. The topology in Theorem 1.8 is called the Skorohod topology, after its inventor. Our proof of Theorem 1.8 will follow Section 3.5 in [EK86]. Let Λ′ be the collection of strictly increasing functions λ mapping [0, ) onto [0, ). ∞ ∞ In particular, for all λ Λ′ we have λ(0) = 0, limt λ(t) = , and λ is continuous. Furthermore,∈ let Λ be the subclass of→∞ Lipschitz continuous∞ functions λ Λ such that ∈ ′ λ(t) λ(s) (1.19) λ := sup log − < . k k t s ∞ 0 s Proof of Lemma 1.9. (i) For all λ Λ, ∈ λ(t) λ(s) λ = sup log − k k 0 s (iii) Since for all λ Λ, ∈ λ λ 1 e−k k k k≥ − log λ(t)−λ(s) = sup 1 e−| t−s | (1.25) 0 s (Counter-)Example. For n N, let ∈ n(n 2) 1 1 n2 −2 t, if t [0, 2 n2 ], − ∈ − 1 1 1 1 1 (1.27) λn(t) := nt + 2 (1 n), if t [ 2 n2 , 2 + n2 ], − ∈ − n(n 2) 2(n 1) 1 1 n2 −2 t n2−2 , t [ 2 + n2 , 1]. − − − ∈ Then (λn)n N in Λ, satisfies (1.22) but λn = n . ∈ k k n−→→∞ ∞ In analogy with (1.16), for v, w E[0, ), we define the Skorohod metric by ∈ D ∞ d dSk(v, w) (1.28) s := inf λ ds e− sup 1 d v t s , w λ(t) s . λ Λ k k∨ ∧ ∧ ∧ ∈ Z[0, ) t [0, ) ∞ ∈ ∞ The next lemma states that dd is indeed a metric on [0, ). Sk DE ∞ Lemma 1.10. [0, ), dd is a metric space. DE ∞ Sk Proof. For symmetry recall Part (i) of Lemma 1.9, and notice that sup 1 d v t s , w λ(t) s t [0, ) ∧ ∧ ∧ (1.29) ∈ ∞ 1 = sup 1 d v λ− (t) s , w t s , t [0, ) ∧ ∧ ∧ ∈ ∞ for all λ Λ. This implies that d (v, w)= d (w, v) for all v, w [0, ). ∈ Sk Sk ∈ DE ∞ If dSk(v, w) = 0, then there exists a sequence (λn)n N in Λ such that ∈ λn 0 and k k n−→→∞ (1.30) ℓ s [0,s0]: sup 1 d v(t s), w(λn(t) s) ε 0 { ∈ t [0, ) ∧ ∧ ∧ ≥ } n−→→∞ ∈ ∞ MARKOV PROCESSES 11 for all ε> 0 and s0 [0, ). Hence by Part (iii) of Lemma 1.9 and (1.30), v(t)= w(t) for all continuity∈ ∞ points t of w, and therefore by Lemma 1.6 and right continuity of v and w, v = w. It remains to show the triangle inequality. Recall Part (ii) of Lemma 1.9, and notice that for all t [0, ), ∈ ∞ sup 1 d w t s , u t λ1 λ2(s) s [0, ) ∧ ∧ ∧ ◦ ∈ ∞ n o sup 1 d w t s , v t λ2(s) ≤ s [0, ) ∧ ∧ ∧ ∈ ∞ n o (1.31) + sup 1 d v t λ2(s) , u t λ1 λ2(s) s [0, ) ∧ ∧ ∧ ◦ ∈ ∞ n o = sup 1 d w t s , v t λ2(s) s [0, ) ∧ ∧ ∧ ∈ ∞ n o + sup 1 d v t s , u t λ1(s) . s [0, ) ∧ ∧ ∧ ∈ ∞ n o Combining (1.24) and (1.31) implies that d (w, u) d (w, v)+d (v, u). Sk ≤ Sk Sk Exercise 1.11. For n N, let vn := 1[0,1 2−n) and wn := 1[0,2−n). Decide ∈ − d whether the sequences (vn)n N and (wn)n N converge in E[0, ), dSk and, if so, determine the limit function.∈ ∈ D ∞ Proposition 1.12 (A convergence criterion). Let (wn) E[0, ) and w [0, ). Then the following are equivalent: ∈ D ∞ ∈ DE ∞ d (a) dSk(wn, w) 0. n−→→∞ (b) There exists a sequence (λn)n N Λ such that λn 0 and ∈ ∈ k k n−→→∞ (1.32) lim sup d wn(λn(t)), w(t) = 0, n →∞ t [0,T ] ∈ for all T [0, ). ∈ ∞ (c) For each T > 0, there exists a sequence (λn)n N in Λ′ (possibly depending on T ) satisfying (1.26) and (1.32). ∈ (d) For each T > 0, there exists a sequence (λn)n N in Λ′ (possibly depending on T ) satisfying (1.17) and (1.18). ∈ Corollary 1.13. The Skorohod topology does not depend on the choice of the metric on (E, ). O 12 JAN SWART AND ANITA WINTER ˜ d Proof of Corollary 1.13. If d, d are two equivalent metrics on (E, ) and dSk d˜ O and dSk are the associated Skorohod metrics, then formula (1.18) shows that d˜ wn w in dSk if and only if wn w in dSk. It is easy to see that two n−→→∞ n−→→∞ metrics are equivalent if every sequence that converges in one metric also converges in the other metric, and vice versa.6 Proof of Proposition 1.12. (a) (b). We start showing that (a) is equiva- ⇐⇒d lent to (b). Assume first that dSk(wn, w) 0 for a metric d on (E, ). By n−→→∞ O definition, then there exist sequences (λn)n N in Λ such that λn 0 ∈ k k n−→→∞ and (1.33) ℓ s [0,s0]: sup 1 d wn(λn(t) s), w(t s) ε 0 { ∈ t [0, ) ∧ ∧ ∧ ≥ } n−→→∞ ∈ ∞ for all ε> 0 and s [0, ). 0 ∈ ∞ Hence, there is a subsequence (nk)k N such that d wn (λn (t) s), w(t ∈ k k ∧ ∧ s) 0 for almost every s [0, ), and thus for all continuity points s of k−→→∞ ∈ ∞ w. That is, there exist sequences (λn)n N in Λ and (sn)n N in [0, ) ∈ ∈ ↑ ∞ ∞ such that λn 0 and k k n−→→∞ (1.34) lim sup d wn(λn(t) sn), w(t sn) = 0. n →∞ t 0 ∧ ∧ ≥ Now for given T [0, ), sn T λn(T ) for all n sufficiently large. There- fore (1.34) implies∈ (1.32).∞ ≥ ∨ On the other hand, let a sequence (λn)n N in Λ satisfy the condition of (b). Let s [0, ). Then for each n N, ∈ ∈ ∞ ∈ sup d wn(λn(t) s), w(t s) t 0 ∧ ∧ ≥ 1 = sup d wn(tn′ s), w(λn− (tn′ ) s) ′ t :=λn(t) 0 ∧ ∧ n ≥ (1.35) 1 sup d wn(tn′ s), w(λn− (tn′ s)) ≤ t′ 0 ∧ ∧ n≥ 1 1 + sup d w(λn− (tn′ s)), w(λn− (tn′ ) s) . ′ tn 0 ∧ ∧ ≥ 6To see this, note that a set A is closed in the topology generated by a metric d if and only if x ∈ A for all xn ∈ A with xn → x in d. This shows that two metrics which define the same form of convergence have the same closed sets. Since open sets are the complements of closed sets, they also have the same open sets, i.e., they generate the same topology. MARKOV PROCESSES 13 We can estimate this further by sup d wn(λn(r)), w(r) ≤ −1 ′ −1 r:=λn (tn s) [0,λn (s)] ∧ ∈ + sup d w(r), w(s) r:=λ−1(t′ ) [s,λ−1(s) s] ∨ n n ∈ n ∨ 1 sup d w(λn− (s)), w(r) , −1 ′ −1 r:=λn (tn) s [λn (s) s,s] ∧ ∈ ∧ where the second half of the last inequality follows by considering the cases t s and t >s separately. Thus by (1.32), n′ ≤ n′ (1.36) lim sup 1 d wn(λn(t) s), w(t s) = 0 n →∞ t [0, ) ∧ ∧ ∧ ∈ ∞ for every continuity point s of w. Hence, applying the dominated conver- d gence theorem in (1.28) yields that dSk wn, w 0. n−→→∞ (b) (c). Obviously, assumption (c) is weaker than (b) (recall also ⇐⇒ N (1.26)). To see the other direction, let N be a positive integer, and (λn )n N ∈ in Λ′ satisfying (1.26) with T = N and such that (1.37) λN (t) := λN (N)+ t N, t N. n n − ≥ We want to construct a sequence (λn)n N in Λ such that ∈ λn 0, and • k k n−→→∞ b supt [0,T ] d wn(λn(t)), w(t) 0, for all T [0, ). • b ∈ n−→→∞ ∈ ∞ Notice that by (1.32) we can find a subsequence (nk)k N such that b ∈ N 1 (1.38) sup d wn(λn (t)), w(t) t [0,N] ≤ N ∈ for all n nN , while in general, we can not conclude from (1.26) that ≥ N lim infn λn = 0 (recall the counterexample given behind the proof of Lemma→∞ 1.9).k k We proceed as follows. First we are therefore going to construct a se- N quence (λn )n N in Λ whose Lipschitz constant goes to 1 as N and ∈ N → ∞ N which is obtained by disturbing (λn )n N such that the dilatation of λn convergesc to zero as N but mildly∈ enough that we can ensure that N → ∞ supt [0,N] d wn(λn (t)), wn(λn(t)) 0. ∈ N−→→∞ For that, define τ N := 0, and for all k 1, c 0 ≥ N N 1 N N inf t>τk 1 : d w(t), w(τk 1) > N , if τk 1 < , (1.39) τk := { − − } N− ∞ , ifτk 1 = . ∞ − ∞ N Since w is right continuous, the sequence (τk )k N is strictly increasing as long as its terms remain finite. Since w has limits from∈ the left, the sequence has no cluster point. Now let for each n N, ∈ N N 1 N (1.40) sk,n := (λn )− (τk ), 14 JAN SWART AND ANITA WINTER where by convention (λN ) 1( )= . n − ∞ ∞ N Define a sequence (λn )n N in Λ by ∈ τ N τ N τ N + k+1− k (t sN ), if t [sN ,sN N), k c sN sN k,n k,n k+1,n k+1,n− k,n − ∈ ∧ N (1.41) λn (t) := N λn (N)+ t N, if t (N, ), − ∈ ∞ c arbitrary, otherwise, c where, by convention 1 = 1. With this convention and by (1.26), ∞− ∞ N N N N N (1.42) λn = max log (τk+1 τk ) log (sk+1,n sk,n) 0, N n→∞ k k sk,n N − − − −→ ≤ and c N 2 (1.43) sup d wn(λn (t)), wn(λn(t)) . t [0,N] ≤ N ∈ Since c (1.44) N sup d wn(λn (t)), w(t) t [0,N] ∈ c N N sup d wn(λn (t)), w(t) + sup d wn(λn (t)), wn(λn(t)) ≤ t [0,N] t [0,N] ∈ ∈ N 2 c sup d wn(λn (t)), w(t) + , ≤ t [0,N] N ∈ for all n N, (1.18) implies that we can choose a subsequence (nk)k N such ∈ ∈ N 1 N 3 that λn N and supt [0,N] d wn(λn (t)), w(t) N for all n nN . For k k≤ ∈ ≤ ≥ N 1 n