Semimartingales and stochastic integration Spring 2011
Sergio Pulido∗
Chris Almost†
Contents
Contents1
0 Motivation3
1 Preliminaries4 1.1 Review of stochastic processes...... 4 1.2 Review of martingales...... 7 1.3 Poisson process and Brownian motion...... 9 1.4 Lévy processes...... 12 1.5 Lévy measures...... 15 1.6 Localization ...... 20 1.7 Integration with respect to processes of finite variation ...... 21 1.8 Naïve stochastic integration is impossible...... 23
∗[email protected] †[email protected]
1 2 Contents
2 Semimartingales and stochastic integration 24 2.1 Introduction...... 24 2.2 Stability properties of semimartingales...... 25 2.3 Elementary examples of semimartingales...... 26 2.4 The stochastic integral as a process...... 27 2.5 Properties of the stochastic integral...... 29 2.6 The quadratic variation of a semimartingale ...... 31 2.7 Itô’s formula...... 35 2.8 Applications of Itô’s formula...... 40
3 The Bichteler-Dellacherie Theorem and its connexions to arbitrage 43 3.1 Introduction...... 43 3.2 Proofs of Theorems 3.1.7 and 3.1.8...... 45 3.3 A short proof of the Doob-Meyer theorem...... 52 3.4 Fundamental theorem of local martingales...... 54 3.5 Quasimartingales, compensators, and the fundamental theorem of local martingales...... 56 3.6 Special semimartingales and another decomposition theorem for local martingales...... 58 3.7 Girsanov’s theorem...... 61
4 General stochastic integration 65 4.1 Stochastic integrals with respect to predictable processes ...... 65
Index 72 Chapter 0
Motivation
Why stochastic integration with respect to semimartingales with jumps? To model “unpredictable” events (e.g. default times in credit risk theory) ◦ one needs to consider models with jumps. A lot of interesting stochastic processes jump, e.g. Poisson process, Lévy ◦ processes. This course will closely follow the textbook, Stochastic integration and differential equations by Philip E. Protter, second edition. We will not cover every chapter, and some proofs given in the course will differ from those in the text. The following numbers correspond to sections in the textbook. I Preliminaries. 1. Filtrations, stochastic processes, stopping times, path regularity, “func- tional” monotone class theorem, optional σ-algebra. 2. Martingales. 3. Poisson processes, Brownian motion. 4. Lévy processes. 6. Localization procedure for stochastic processes. 7. Stieltjes integration 8. Impossibility of naïve stochastic integration (via the Banach-Steinhaus theorem). II Semimartingales and stochastic integrals. 1–3. Definition of the stochastic integral with respect to processes in L. 5. Properties of the stochastic integral. 6. Quadratic variation. 7. Itô’s formula. 8. Stochastic exponential and Lévy’s characterization theorem. III Bichteler-Dellacherie theorem. (NFLVR) implies S is a semimartingale (NFLVR) and little investment if and only if S is a semimartingale IV Stochastic integration with respect to predictable processes and martingale representation theorems (i.e. market completeness). For more information on the history of the development of stochastic integra- tion, see the paper by Protter and Jarrow on that topic.
3 Chapter 1
Preliminaries
1.1 Review of stochastic processes
The standard setup we will use is that of a complete probability space (Ω, , P) F and a filtration F = ( t )0 t of sub-σ-algebras of . The filtration can be thought of as the flowF of information.≤ ≤∞ Expectation E willF always be with respect to P unless stated otherwise. Notation. We will use the convection that t, s, and u will always be real variables, not including unless it is explicitly mentioned, e.g. t t 0 = [0, ). On the other hand, n ∞and k will always be integers, e.g. n : n { |0 ≥= }0, 1, 2, 3,∞ . . . =: N. { ≥ } { } 1.1.1 Definition. (Ω, , F, P) satisfies the usual conditions if F (i) 0 contains all the P-null sets. F T (ii) F is right continuous, i.e. t = s
4 1.1. Review of stochastic processes 5
1.1.4 Theorem. V W (i) If (Tn)n 1 is a sequence of stopping times then n Tn and n Tn are stopping times. ≥ (ii) If T and S are stopping times then T + S is a stopping time.
1.1.5 Exercises. (i) If T S then is T S a stopping time? (ii) For which≥ constants− α is αT a stopping time?
SOLUTION: Clearly αT need not be a stopping time if α < 1. If α 1 then, for any ≥ t 0, t/α t so αT t = T t/α t/α t and αT is a stopping time. ≥Let H be≤ a stopping{ ≤ time} { for≤ which}H ∈/ F2 is not⊆ F a stopping time (e.g. the first hitting time of Brownian motion at the level 1). Take T = 3H/2 and S = H, both stopping times, and note that T S = H/2 is not a stopping time. ú − 1.1.6 Definition. Let T be a stopping time. The σ-algebras of events before time T and events strictly before time T are (i) T = A : A T t t for all t . F { ∈ F ∩ { ≤ } ∈ F } (ii) T = 0 σ A t < T : t [0, ), A t . F − F ∨ { ∩ { } ∈ ∞ ∈ F } 1.1.7 Definition. d (i)A stochastic process X is a collection of R -valued r.v.’s, (X t )0 t< . A stochas- d tic process may also be thought of as a function X : Ω [0, ≤ )∞ R or as a random element of a space of paths. × ∞ →
(ii) X is adapted if X t t for all t. ∈ F (iii) X is càdlàg (X t (ω))t 0 has left limits and is right continuous for almost all ω. X is càglàd if instead≥ the paths have right limits and are left continuous. (iv) X is a modification of Y if P[X t = Yt ] = 0 for all t. X is indistinguishable 6 from Y if P[X t = Yt for some t] = 0. 6 (v) X := (X t )t 0 where X0 := 0 and X t := lims t Xs. − − ≥ − − ↑ (vi) For a càdlàg process X , ∆X := (∆X t )t 0 is the process of jumps of X , where ≥ ∆X t := X t X t . − − 1.1.8 Theorem. (i) If X is a modification of Y and X and Y are left- (right-) continuous then they are indistinguishable. d (ii) If Λ R is open and X is continuous from the right (càd) and adapted then ⊆ T := inf t > 0 : X t Λ is a stopping time. {d ∈ } (iii) If Λ R is closed and X is càd and adapted then T := inf t > 0 : X t ⊆ { ∈ Λ or X t Λ is a stopping time. − ∈ } (iv) If X is càdlàg and adapted and ∆X T 1T< = 0 for all stopping times T then ∆X is indistinguishable from the zero process.∞
PROOF: Read the proofs of these facts as an exercise. 6 Preliminaries
1.1.9 Definition. = σ(X : X is adapted and càdlàg) is the optional σ-algebra. A stochastic processO X is an optional process if X is -measurable. O 1.1.10 Theorem (Début theorem). If A then T(ω) := inf t : (ω, t) A , the début time of A, is a stopping time. ∈ O { ∈ }
Remark. This theorem requires that the filtration is right continuous. For example, suppose that T is the hitting time of an open set by a left continuous process. Then
you can prove T < t t for all t without using right continuity of the filtration, { } ∈ F but you cannot necessarily prove that T = t t without it. You need to “look into the future” a little bit. { } ∈ F
d 1.1.11 Corollary. If X is optional and B R is a Borel set then the hitting time ⊆ T := inf t > 0 : X t B is a stopping time. { ∈ } 1.1.12 Theorem. If X is an optional process then (i) X is ( ([0, )))-measurable and (ii) X T 1T 1.1.13 Theorem (Monotone class theorem). Suppose that H is collection of bounded R-valued functions such that (i) H is a vector space. (ii) 1Ω H, i.e. constant functions are in H. (iii) If (∈fn)n 0 H is monotone increasing and f := limn fn (pointwise) is bounded≥ then⊆ f H. →∞ (In this case H is called∈ a monotone vector space.) Let M be a multiplicative col- lection of bounded functions (i.e. if f , g M then f g M). If M H then H contains all the bounded functions that are∈ measurable with∈ respect to⊆ σ(M). PROOF (OF THEOREM 1.1.12): Use the Monotone Class Theorem. Define M := X : Ω [0, ) R X is càdlàg, adapted, and bounded { × ∞ → | } H := X : Ω [0, ) R X is bounded and (i) and (ii) hold { × ∞ → | } It can be checked that H is a monotone vector space, M is a multiplicative collec- tion, and σ(M) = . If we prove that M H then we are done. Let X M and define O ⊆ ∈ (n) X∞ X : X k 1 k 1 , k . = 2n [ 2−n 2n ) k=1 Since X is right continuous, X (n) X pointwise. Let B be a Borel set. → n o k 1 k (n) [∞ X B = X k B −n , n ([0, )) 2n 2 2 { ∈ } k=1 ∈ × ∈ F ⊗ B ∞ 1.2. Review of martingales 7 This proves that X satisfies (i). To prove (ii), let T be a stopping time and define ¨ k k 1 k if T < 2n 2−n 2n Tn := ≤ if T = . ∞ ∞ Then the Tn’s are stopping times and Tn T. ↓ [∞ n o k X B T t X k B T Tn n = n = n t 2n 2 { ∈ } ∩ { ≤ } k=1 ∈ ∩ ∈ F k t 2n ≤ This shows that X . Since X is right continuous, Tn Tn ∈ F X T 1T< lim X T 1T < . = n n n ∞ ∞ →∞ It can be shown that, since the filtration is right continuous, T , so ∞n 1 Tn = T = F F X T 1T< T and X satisfies (ii). Therefore X H. ∞ ∈ F ∈ If X is càdlàg and adapted then an argument similar to that in the previous proof shows that X Ω [0,t] t ([0, t]) for all t, i.e. X is progressively measur- able. By a similar monotone| × ∈ F class⊗ B argument, it can be shown that every optional process is a progressive process. 1.1.14 Definition. := σ(X : X is progressively measurable) is the progressive σ-algebra. V 1.1.15 Corollary. . O ⊆ V 1.2 Review of martingales 1.2.1 Theorem. Let X be a (sub-, super-) martingale and assume that F satisfies the usual conditions. Then X has a right continuous modification if and only if the function t E[X t ] is right continuous. Furthermore, this modification has left limits everywhere.7→ PROOF (SKETCH): The process Xet := lim s t Xs is the correct modification. s ↓Q ∈ 1.2.2 Corollary. Every martingale has a càdlàg modification, unique up to indis- tinguishability. + 1.2.3 Theorem. Let X be a right continuous sub-martingale with supt 0 E[X t ] < 1 ≥ . Then X := limt X t exists a.s. and X L . ∞ ∞ →∞ ∞ ∈ 1.2.4 Definition. A collection of random variables (Uα)α A is uniformly integrable or u.i. if ∈ lim sup 1 U >nUα 0. n E[ α ] = α A | | →∞ ∈ 8 Preliminaries 1.2.5 Theorem. The following are equivalent for a family (Uα)α A. ∈ (i) (Uα)α A is u.i. ∈ (ii) supα A E[ Uα ] < and for all " > 0 there is δ > 0 such that if P[Λ] < δ ∈ | | ∞ then supα A E[1ΛUα] < ". (iii) (de la Vallée-Poussin∈ criterion) There is a positive, increasing, convex func- tion G on 0, such that lim G(x) and sup G U < . [ ) x x = α A E[ ( α )] ∞ →∞ ∞ ∈ | | ∞ 1.2.6 Theorem. The following are equivalent for a martingale X . (i) (X t )t 0 is u.i. ≥ (ii) X is closable, i.e. there is an integrable r.v. Z such that X t = E[Z t ]. (iii) X converges in L1. |F 1 (iv) X = limt X t a.s. and in L and X closes X . ∞ →∞ ∞ Remark. (i) If X is u.i. then X is bounded in L1 (but not vice versa), so Theorem 1.2.3 implies that X t X a.s. Theorem 1.2.6 upgrades this to convergence in 1 → ∞ L , i.e. E[ X t X ] 0. | − ∞| → W (ii) If X is closed by Z then E[Z ] also closes X , where := t 0 t . Furthermore, X = E[Z ].|F∞ F∞ ≥ F ∞ |F∞ 1.2.7 Example (Simple random walk). Let (Zn)n 1 be i.i.d. with ≥ 1 P[Zn = 1] = P[Zn = 1] = . − 2 P t Take t := σ Zk : k t and X t := bk c1 Zk. Then X is not a closable martingale. F { ≤ } = 1.2.8 Example. Let M : exp B 1 t , the stochastic exponential of Brownian t = ( t 2 ) motion, a martingale. Then Mt 0− a.s. by Theorem 1.2.3 because it is a positive valued martingale (hence M →is a sub-martingale with no positive part). But − E[ Mt ] = 1 for all t so M is not u.i. by Theorem 1.2.6. | | 1.2.9 Theorem. (i) If X is a closable (sub-, super-) martingale and S T are stopping times ≤ then E[X T S] = XS ( , ). (ii) If X is a (sub-,|F super-)≥ martingale≤ and S T are bounded stopping times ≤ then E[X T S] = XS ( , ). (iii) If X is a right|F continuous≥ ≤ sub-martingale and p > 1 then p sup X t Lp sup X t Lp . k t 0 k ≤ p 1 t 0 k k ≥ − ≥ 2 2 In particular if p = 2 then E[supt 0 X t ] 4 supt 0 E[X t ]. (iv) (Jensen’s inequality) If ϕ is a convex≥ function,≤ Z≥is an integrable r.v., and is a sub-σ-algebra of then ϕ(E[Z ]) E[ϕ(Z) ]. G F |G ≤ |G 1.3. Poisson process and Brownian motion 9 1.2.10 Definition. Let X be a process and T be a random time. The stopped T process is X t := X t 1t T + X T 1t>T,T< = X T t . ≤ ∞ ∧ If T is a stopping time and X is càdlàg and adapted then so is X T . 1.2.11 Theorem. (i) If X is a u.i. martingale and T is a stopping time then X T is a u.i. martingale. (ii) If X is a martingale and T is a stopping time then X T is a martingale. That is, martingales are “stable under stopping”. 1.2.12 Definition. Let X be a martingale. 2 (i) If X t L for all t then X is called a square integrable martingale. ∈ 2 2 (ii) If (X t )t [0, ) is u.i. and X L then X is called an L -martingale. ∈ ∞ ∞ ∈ 1.2.13 Exercise. Any L2-martingale is a square integrable martingale, but not conversely. 2 2 SOLUTION: Let X be an L -martingale. Then X L and so, by the conditional version of Jensen’s inequality, ∞ ∈ 2 2 2 2 E[X t ] = E[(E[X t ]) ] E[E[X t ]] = E[X ] < . ∞|F ≤ ∞|F ∞ ∞ 2 Therefore X t L for all t. We have already seen that the stochastic exponential of Brownian motion∈ is not u.i. (and hence not an L2-martingale) but it is square integrable because the normal distribution has a finite valued moment generating function. ú 1.3 Poisson process and Brownian motion 1.3.1 Definition. Suppose that (Tn)n 1 is a strictly increasing sequence of ran- dom times with T > 0 a.s. The process≥ N P 1 is the counting process 1 t = n 1 Tn t ≥ ≤ associated with (Tn)n 1. The random time T := supn 1 Tn is the explosion time. If T = a.s. then N is≥ a counting process without explosion.≥ ∞ 1.3.2 Theorem. A counting process is an adapted process if and only if Tn is a stopping time for all n. PROOF: If N is adapted then t < Tn = Nt < n t for all t and all n. There- fore, for all n, Tn t t{for all }t, so{ Tn is a} stopping ∈ F time. Conversely, if all { ≤ } ∈ F the Tn are stopping times then, for all t, Nt n = t Tn t for all n. Since { ≤ } { ≤ } ∈ F N takes only integer values this implies that Nt t . ∈ F 1.3.3 Definition. An adapted process N is called a Poisson process if () N is a counting process. (i) Nt Ns is independent of s for all 0 s < t < (independent increments). − F ≤ ∞ 10 Preliminaries (d) (ii) Nt Ns = Nt s for all 0 s < t < (stationary increments). − − ≤ ∞ Remark. Implicit in this definition is that a Poisson process does not explode. The definition can be modified slightly to allow this possibility, and then it can be proved as a theorem that a Poisson process does not explode, but the details are very technical and relatively unenlightening. 1.3.4 Theorem. Suppose that N is a Poisson process. (i) N is continuous in probability. (d) (ii) Nt = Poisson(λt) for some λ 0, called the intensity or arrival rate of N. In particular, Nt has finite moments≥ of all orders for all t. 2 (iii) (Nt λt)t 0 and ((Nt λt) λt)t 0 are martingales. −N ≥ − N− ≥N N (iv) If t := σ(Ns : s t) and F = ( t )t 0 then F is right continuous. F ≤ F ≥ PROOF: Let α(t) := P[Nt = 0] for t 0. For all s < t, ≥ α(t + s) = P[Nt+s = 0] = P[ Ns = 0 Nt+s Ns = 0 ] non-decreasing and non-negative { } ∩ { − } = P[Ns = 0] P[Nt+s Ns = 0] independent increments − = P[Ns = 0] P[Nt = 0] stationary increments = α(t)α(s) If t t then N 0 N 0 , so α is right continuous and decreasing. It n tn = t = ↓ { }%{ } λt follows that either α 0 or α(t) = e− for some λ 0. By the definition of ≡ ≥ counting process N0 = 0, so α is cannot be the zero function. (i) Observe that, given " > 0 small, for all s < t, P[ Nt Ns > "] = P[ Nt s > "] stationary increments | − | | − | = P[Nt s > "] N is non-decreasing − = 1 P[Nt s = 0] N is integer valued − − λ(t s) = 1 e− − 0− as s t. → → Therefore N is left continuous in probability. The proof of continuity from the right is similar. (ii) First we need to prove that lim 1 N 1 λ. Towards this, let t 0 t P[ t = ] = → β(t) := P[Nt 2] for t 0. If we can show that limt 0 β(t)/t = 0 then we would have ≥ ≥ → P[Nt = 1] 1 α(t) β(t) lim = lim − − = λ. t 0 t t 0 t → → It is enough to prove that limn nβ(1/n) = 0. Divide [0, 1] into n equal →∞ subintervals and let Sn be the number of subintervals with at least two ar- (d) rivals. It can be see that Sn = Binomial(n, β(1/n)) because of the stationary and independent increments of N. In the definition of counting process the 1.3. Poisson process and Brownian motion 11 sequence of jump times is strictly increasing, so limn Sn = 0 a.s. Clearly →∞ Sn < N1, so limn nβ(1/n) = limn E[Sn] = 0 provided that E[N1] < . We are going to→∞ gloss over this point.→∞ ∞ N Let ϕ(t) := E[γ t ] for 0 < γ < 1. Nt+s Ns Nt+s Ns Ns Nt ϕ(t + s) = E[γ ] = E[γ γ − ] = E[γ ] E[γ ] = ϕ(t)ϕ(s) and ϕ is right continuous because N has right continuous paths. Therefore tψ γ ϕ(t) = e ( ) for some function ψ of γ. By definition, X n ϕ(t) = γ P[Nt = n] n 0 ≥ tψ(γ) X n e = α(t) + γ P[Nt = 1] + γ P[Nt = n]. n 2 ≥ Differentiating, ψ(γ) = limt 0(ϕ(t) 1)/t = λ + γλ by the computations → n above. Comparing coefficients of γ −in the power− series shows that Nt has a λt n Poisson distribution with rate λt, i.e. P[Nt = n] = e− (λt) /n!. (iii) Exercise. (iv) See the textbook. 1.3.5 Definition. An adapted process B is called a Brownian motion if (i) Bt Bs is independent of s for all 0 s < t < (independent increments). − (d) F ≤ ∞ (ii) Bt Bs = Normal(0, t s) for all 0 s < t < (stationary, normally distributed,− increments).− ≤ ∞ If B0 0 then B is a standard Brownian motion. ≡ 1.3.6 Theorem. Let B be a Brownian motion. (i) If E[ B0 ] < then B is a martingale. (ii) There| is| a modification∞ of B with continuous paths. 2 (iii) (Bt t)t 0 is a martingale when B is standard BM. − ≥ (iv) Let Πn be a refining sequence of partitions of the interval [a, a + t] with P 2 2 limn mesh n 0. Then nB : Bt Bt t in L and a.s. (Π ) = Π = tn Πn ( i+1 i ) when→∞B is standard BM. ∈ − → (v) For almost all ω, the function t Bt (ω) is of unbounded variation. 7→ P Remark. To prove (ii) with the additional assumption that n 0 mesh(Πn) < , but dropping the assumption that the partitions are refining, you≥ can use the Borel-∞ Cantelli lemma. The proof of (iv) uses the backwards martingale convergence theorem. 12 Preliminaries 1.4 Lévy processes 1.4.1 Definition. An adapted, real-valued process X is called a Lévy process if () X0 = 0 (i) X t Xs is independent of s for all 0 s < t < (independent increments). − (d) F ≤ ∞ (ii) X t Xs = X t s for all 0 s < t < (stationary increments). − − ≤ ∞ (iii) (X t )t 0 is continuous in probability. ≥ If only (i) and (iii) hold then X is an additive process. If (X t )t 0 is non-decreasing then X is called a subordinator. ≥ X Remark. If the filtration is not specified then we assume that F = F is the natural (not necessarily complete) filtration of X . In this case X might then be called an intrinsic Lévy process. 1.4.2 Example. The Poisson process and Brownian motion are both Lévy pro- cesses. Let W be a standard BM and define Tb := inf t > 0 : Wt b . Then T is an intrinsic Lévy process and subordinator for{ the filtration≥ } . ( b)b 0 ( Tb )b 0 Indeed,≥ if 0 a < b < then T T is independent of and distributedF ≥ as a b Ta Tb a by the≤ strong Markov∞ property− of W. We will see thatF T is a stable process with− parameter α = 1/2. izX t tΨ(z) 1.4.3 Theorem. Let X be a Lévy process. Then ft (z) := E[e ] = e− for z izX t some continuous function Ψ = ΨX . Furthermore, Mt := e /ft (z) is a martingale for all z R. ∈ PROOF: Fix z R. By the stationarity and independence of the increments, ∈ ft (z) = ft s(z)fs(z) for all 0 s < t < . (1.1) − ≤ ∞ We would like to show that t ft (z) is right-continuous. By the multiplicative 7→ property (1.1), it suffices to show t ft (z) is right continuous at zero. Suppose that t 0; we need to show that f 7→z 1. By definition, f z 1, so we are n tn ( ) tn ( ) done if↓ we can show that every convergent→ subsequence converges| | ≤ to 1. Suppose without loss of generality that f z a . Since X is continuous in probability, tn ( ) C izX e tn 1 in probability. Along a subsequence→ ∈ we have convergence almost surely, i.e. f → z 1. Therefore a 1 and we are done. tt ( ) = nk → tΨ(z) The multiplicative property and right continuity imply that ft (z) = e− for some number Ψ(z). Since ft (z) is a characteristic function, z ft (z) is contin- uous (use dominated convergence). Therefore Ψ must be continuous7→ as well. In particular, ft (z) = 0 for all t and all z. Let 0 s < t < . 6 ≤ ∞ eizX t eizXs iz(X t Xs) E s = E[e − s] independent increments ft (z) F ft (z) |F eizXs = ft s(z) stationary increments ft (z) − 1.4. Lévy processes 13 eizXs = f is multiplicative. fs(z) z Therefore Mt is a martingale. 1.4.4 Theorem. If X is an additive process then X is Markov with transition func- tion Ps,t (x, B) = P[X t Xs B x]. In particular, X is spatially homogeneous, i.e. − ∈ − Ps,t (x, B) = Ps,t (0, B x). − 1.4.5 Corollary. If X is additive, " > 0, and T < then limu 0 α",T (u) = 0, where ↓ α",T is defined in the next theorem. ∞ C PROOF: Ps,t (x, B"(x) ) = P( X t Xs ") 0 as s t uniformly on [0, T] in probability. It is an exercise to| show− that| ≥ the→ continuity→ in probability of X implies uniform continuity in probability on compact intervals. d 1.4.6 Theorem. Let X be a Markov process on R with transition function Ps,t . If limu 0 α",T (u) = 0, where ↓ C d α",T (u) = sup Ps,t (x, B"(x) ) : x R , s, t [0, T], 0 t s u , { ∈ ∈ ≤ − ≤ } then X has a càdlàg modification. If furthermore limu 0 α",T (u)/u = 0 then X has a continuous modification. ↓ PROOF: See Lévy processes and infinitely divisible distributions by Kin-Iti Sato. The important steps are as follows. Fix " > 0 and ω Ω. Say that X (ω) has an "-oscillation n-times in M [0, ) if there are t < t ∈< < t all in M such that X ω X ω "⊆. X has∞ 0 1 n t j ( ) t j 1 ( ) "-oscillation infinitely often··· in M if this holds for all| n. Define− − | ≥ \∞ \∞ : ω : X ω does not have 1 -oscillation infinitely often in 0, N Ω20 = ( ) k [ ] Q N=1 k=1{ ∩ } It can be shown that Ω02 , and also that ∈ F n o Ω20 ω : lim Xs exists for all t and lim Xs exists for all t > 0 . s t,s Q s t,s Q ⊆ ↓ ∈ ↑ ∈ The hard part is to show that P[Ω02] = 1. Remark. (i) If X is a Feller process then X has a càdlàg modification (see Revuz-Yor). (ii) From now on we assume that we are working with a càdlàg version of any Lévy process that appears. (iii) P[X t = X t ] = 0 for any process X that is continuous in probability. Of course,6 this− does not mean that X doesn’t jump, it means that X has no fixed times of discontinuity. 14 Preliminaries 1.4.7 Theorem. Let X be a Lévy process and be the collection of P-null sets. If X N t := t for all t then G = ( t )t 0 is right continuous. G F ∨ N G ≥ n n ROOF P : Let (s1,..., sn) and (u1,..., un) be arbitrary vectors in (R+) and R re- spectively. We first want to show that i u X u X i u X u X ( 1 s1 + + n sn ) ( 1 s1 + + n sn ) E[e ··· t ] = E[e ··· t+]. |G |G It is clear that it suffices to show this with s1,..., sn > t. We take n = 2 for notational simplicity, and assume that s2 s1 > t. ≥ i(u1 Xs +u2 Xs ) i(u1 Xs +u2 Xs ) e 1 2 t lim e 1 2 w exercise E[ +] = w t E[ ] |G ↓ |G iu1 Xs u2 lim e 1 M w fs u2 = w t E[ s2 ] 2 ( ) ↓ |G iu1 Xs u2 lim e 1 M w fs u2 tower property = w t E[ s1 ] 2 ( ) ↓ |G i u u X lim e ( 1+ 2) s1 f u f is multiplicative = E[ w] s2 s1 ( 2) w t − ↓ |G ei(u1+u2)Xw lim f u u f u M is a martingale = s1 ( 1 + 2) s2 s1 ( 2) w t fw(u1 + u2) − ↓ ei(u1+u2)X t f u u f u X is càdlàg = s1 t ( 1 + 2) s2 s1 ( 2) − − By the same steps, we obtain i u X u X i u u X e ( 1 s1 + 2 s2 ) e ( 1+ 2) t f u u f u . E[ t ] = s1 t ( 1 + 2) s2 s1 ( 2) |G − − X By a monotone class argument, E[Z t ] = E[Z t+] for any bounded Z . It |G |G ∈ F∞ follows that t+ t consists only of sets from . Since t , they must be equal. G \G N N ⊆ G T X 1.4.8 Corollary (Blumenthal 0-1 Law). Let X be a Lévy process. If A t>0 t then P[A] = 0 or 1. ∈ F 1.4.9 Theorem. If X is a Lévy process and T is a stopping time then, on T < , { ∞} Yt := X T+t X T is a Lévy process with respect to ( t+T )t 0 and Y has the same finite dimensional− distributions as X . F ≥ 1.4.10 Corollary. If X is a Lévy process then X is a strong Markov process. n 1.4.11 Theorem. If X is a Lévy process with bounded jumps then E[ X t ] < for all t and all n. | | ∞ PROOF: Suppose that supt ∆X t C. Define a sequence of random times Tn recursively as follows. T0 |0 and| ≤ ≡ ¨ inf t > T : X X C if T < T n t Tn n n+1 = { | − | ≥ } ∞ if Tn = ∞ ∞ 1.5. Lévy measures 15 Tn By definition of the Tn, sups Xs 2nC. The 2 comes from the possibility that X could jump by C just before| hitting| ≤ the stopping level. Since X is càdlàg, the Tn are all stopping times. Further, since X is a strong Markov process on Tn < , (i) T T is independent of and { ∞} n n 1 Tn 1 − − (d) F − (ii) Tn Tn 1 = T1. Therefore− − n Y Tn (Tk Tk 1) T1 n n E[e− ] = E e− − − (E[e− ]) =: α , k=0 ≤ where 0 α < 1. The comes from the fact that e−∞ = 0 and some of the Tn may be . (We≤ interpret T≤k Tk 1 to be if both of them are .) By Chebyshev’s inequality,∞ − − ∞ ∞ Tn E[e− ] X > 2nC T < t αnet . P[ t ] P[ n ] e t | | ≤ ≤ − ≤ Finally, X∞ eβ X t 1 eβ X t 1 E[ | |] + E[ | | 2nC< X t 2(n+1)C ] | |≤ ≤ n=0 X∞ 2β(n+1)C 1 + e ) P[ X t > 2nC] ≤ n=0 | | 2βC t X∞ 2βC n 1 + e e (αe ) ≤ n=0 Choosing an appropriately small, positive β shows that X t has an exponential moment, so it has polynomial moments of all orders. | | 1.5 Lévy measures Suppose that Λ (R) and 0 / Λ. Let X be a Lévy process (with càdlàg paths, as always) and define∈ B inductively∈ a sequence of random times T 0 0 and Λ ≡ n+1 n T := inf t > T : ∆X t Λ . Λ { Λ ∈ } These times have the following properties. (i) T n t by the usual conditions and T n is an increasing Λ t+ = t ( Λ )n 1 sequence{ ≥ } ∈of F (possiblyF -valued) stopping times. ≥ (ii) T 1 > 0 a.s. since X ∞0, X has càdlàg paths, and 0 / . Λ 0 Λ (iii) lim T n since≡ X has càdlàg paths and 0 ∈/ (see Homework 1, n Λ = Λ problem→∞ 10). ∞ ∈ Let N Λ be the number of jumps with size in Λ before time t. X X∞ Λ n Nt := 1Λ(∆Xs) = 1T t Λ ≥ 0 Since X is a strong Markov process, N Λ has stationary and independent incre- ments. It is also a counting process with no explosions, so it is a Poisson process. Λ Λ By Theorem 1.1.12(i) we can define ν(Λ) := E[N1 ], the intensity of N . We can extend this definition to the case when 0 Λ, so long as 0 / Λ by taking ν(Λ) = P ∈ ∈ ∞ whenever 0 1.5.3 Example. If X is a Poisson process of rate λ then ν = λδ1. 1.5.4 Theorem. If X is a Lévy process then it can be decomposed as X = Y + Z T p where Y is a Lévy process and martingale with bounded jumps (so Yt p 1 L for all t) and Z is a Lévy process with paths of finite variation on compact∈ subsets≥ of [0, ). ∞ 1.5.5 Theorem. Let X be a Lévy process with jumps bounded by C. The process c d Zt := X t E[X t ] is a martingale that can be decomposed as Z = Z + Z , where Z c and Z−d are independent Lévy processes. Moreover, Z d Zt := x(N(t, d x) tν(d x)), x C − | |≤ c where N(t, d x) is the measure of Theorem 1.5.1, and the remainder Z has con- tinuous paths. 1.5.6 Lemma. If X is a subordinator with continuous paths then X t = ct for some constant c. ROOF P : X 0(0) = lim" 0 X"/" exists a.s., and since X 0(0) 0, it is a constant a.s. ↓ ∈ F Let δ > 0 and inductively define a sequence of stopping times by T0 = 0 and T : inf t > T : X X c δ t . k+1 = k t Tk ( + ) { − ≥ } Then, since X is continuous, X c δ T and X c δ t for t T . T1 = ( + ) 1 t ( + ) 1 ≤ ≤ Additionally, X t (c + δ)t for t Tk. Because X is a Lévy process, (Tk+1 Tk)k 0 are strictly positive≤ i.i.d. random≤ variables. By the strong law of large numbers− ≥ Tk as k , so X t (c + δ)t for all t. Therefore X t ct for all t, and the → ∞ → ∞ ≤ ≤ proof that X t ct for all t is similar. ≥ 1.5.7 Theorem (Lévy-Itô decomposition). If X is a Lévy process then there are constants σ and α such that Z (d) X X σB αt x N t, d x tν d x X 1 t = t + + ( ( ) ( )) + ∆ s ∆Xs 1 x <1 − 0 Remark. X has finite variation on compacts if either (i) σ 0 and ν < or = (R) R (ii) σ 0 and ν ∞ but x ν d x < . = (R) = x <1 ( ) R ∞ | | ∞ If σ 0 or x ν d x | then| X has infinite variation on compacts. To = x <1 ( ) = prove6 that the| continuous| | | part∞ of X is σB we will need Lévy’s characterization of BM, proven later. 1.5.8 Lemma. Let X be a Lévy process, Λ R 0 be Borel with 0 / Λ, and let f be Borel measurable and finite valued on⊆Λ. \{ } ∈ (i) The process Y defined by Z X Y : f x N t, d x f X 1 t = ( ) ( ) = (∆ s) ∆Xs Λ Λ 0 X t = X t Jt αt + Jt + αt . | − {z− } | {z } Yt Zt PROOF (OF THEOREM 1.5.5): That Zt := X t E[X t ] is a martingale is an exercise. Suppose without loss of generality that the− jumps of X are bounded by 1. Define 1 1 R R : < x and M Λk : xN t, d x t xν d x . Then the M Λk are Λk = k 1 k t = ( ) ( ) { + | | ≤ } Λk − Λk martingales by lemma (ii) and are in L2 by lemma (iii) and they are independent n Pn by lemma (iv). Let M : M Λk . = k=1 n n Z Z n X X 2 2 Var(M ) = Var(M Λk ) = t x ν(d x) = t x ν(d x) 1 < x 1 k=1 k=1 Λk n 1 + | |≤ 18 Preliminaries Prove as an exercise that Z M n is independent of M n (it is not an immediate − n n consequence of the lemma). Therefore Var(M ) = Var(Zt ) Var(Z M ) < − − ∞ since Var(Zt ) < because Z has bounded jumps. Since M n is∞ a sum of independent random variables we can take the limit as 2 c n n in L (fill in the details as an exercise). Let Z := limn (Z M ) and d→ ∞ n c d n →∞ − Z := limn M . Then Z is independent of Z since Z M is independent of M n for all n→∞. The formula for Z d is immediate. − 1 Something is missing from the last To show Z c is continuous, note first that Z M n has jumps bounded by . n+1 part of this proof. n− 2 By Doob’s maximal inequality sup0 σ2z2 Z z iαz 1 eizx izx1 ν d x . Ψ( ) = 2 + ( + x <1) ( ) − R − | | PROOF (SKETCH): In the Lévy-Itô decomposition it can be shown that the parts are all independent, so the characteristic functions multiply. Show as an exercise that the characteristic function of R X t eizx 1 ν d x X 1 is e x 1( ) ( ) ∆ s ∆Xs 1 | |≥ − 0 Z R t 1 eizx izx ν d x x <1( + ) ( ) x(N(t, d x) tν(d x)) is e | | − . x <1 − | | 1.5.10 Definition. A probability distribution F on R is an infinitely divisible distri- (d) bution if for all n there are X1,..., X n i.i.d. such that X1 + + X n = F. ··· 1.5.11 Theorem. (i) If X is a Lévy process then, for all t, X t has an infinitely divisible distribution. (ii) Conversely, if F is an infinitely divisible distribution then there is a Lévy (d) process X such that X1 = F. (iii) If F is an infinitely divisible distribution then its characteristic function has the form of the Lévy-Khintchine exponent with t 1, where ν is a measure R = on such that ν 0 0 and x 2 1 ν d x < , and the representa- R ( ) = R( ) ( ) tion of F in terms{ of}(σ, α, ν) is unique.| | ∧ ∞ 1.5.12 Examples. iz λt(e 1) (i) If X Poisson(λ) then the characteristic function is e − , so σ = α = 0 ∼ and ν = λδ1. 1.5. Lévy measures 19 (ii) The Γ-process is the Lévy process such that X1 has a Γ-distribution, where bc c 1 bx X1 d x x e 1x>0d x for some constants b, c > 0. The Lévy- P[ ] = Γ(c) − Khintchine∈ exponent is Z e bx c eizx 1 − 1 d x. ( ) x x>0 R − In this case ν d x e bx /x1 d x, σ 0, and α c 1 e b > 0. The ( ) = − x>0 = = b ( − ) paths of X are non-decreasing since it has positive drift, no− volatility, and positive jumps. Because ν((0, 1)) = , X has infinite activity. (iii) Stable processes are those with Lévy-Khintchine∞ exponent of the form Z ∞ izx izx 1 izδ m1 e 1 d x + 2 1+γ 0 − − 1 + x x Z 0 izx izx 1 + m2 e 1 2 1 γ d x − − 1 + x x + −∞ where 0 < γ < 2. When m m one can take ν d x 1 d x. In this 1 = 2 ( ) = x 1+γ | | case there are Y1, Y2, . . . i.i.d. and constants an, and bn such that 1 X∞ (w) Yi bn X1 an i=1 − −→ 1 (d) γ R and (β − Xβ t )t 0 = (X t )t 0 for all β > 0. If 1 α < 2 then x ν(d x) = so the process would≥ have≥ infinite variation on≤ compacts. | | ∞ (iv) For the hitting time process (Tb)b 0 of standard Brownian motion B, we have ≥ b b2 T d t e 2t 1 d t P[ b ] = p − t>0 ∈ 2πt3 so the Lévy-Khintchine exponent is Z b ∞ izx 3 2 (e 1)x − d x. p2π 0 − It follows that T is a stable process with γ 1 . Indeed, = 2 1 1 2 Tβ b = 2 inf t : Bt = β b β β { } 2 = inf t/β : Bt /β = b { (d)} = inf t : Bβ 2 t /β = b = Tb { } (d) since (Bβ 2 t /β)t 0 = B. Also note that T is non-decreasing. ≥ 20 Preliminaries 1.6 Localization 1.6.1 Definition. (i) If C is a collection of processes then the localized class, Cloc, is the collection of processes X such that there exists a sequence (Tn)n 1 of stopping times Tn ≥ such that Tn a.s. and X C for all n. The sequence (Tn)n 1 is called a localizing sequence↑ ∞ for X relative∈ to C. ≥ (ii) If C is the collection of adapted, càdlàg, uniformly integrable martingales then an element of the localized class is called a local martingale. (iii) A stopping time T reduces X relative to C if X T C. (iv) C is stable under stopping if X T C for all X C∈and all stopping times T. ∈ ∈ 1.6.2 Theorem. Suppose that C is a collection of processes that is stable under stopping. (i) Cloc is stable under stopping and (Cloc)loc = Cloc. In particular, a local local martingale is a local martingale. (ii) If T reduces X relative to C and S T a.s. then S reduces X . (iii) If M and N are local martingales then≤ M + N is a local martingale. If S and T reduce M then S T reduces M. ∨ (iv) (C C 0)loc = Cloc C 0loc. ∩ ∩ PROOF (OF (I)): Let X Cloc and T be a stopping time. If (Tn)n 1 is a localizing T T T T sequence for X then (X∈ ) n = (X n ) C since C is stable under≥ stopping. There- ∈ T fore (Tn)n 1 is a localizing sequence for X and it is seen that Cloc is stable under stopping. ≥ Now let X (Cloc)loc and (Tn) be a localizing sequence for X relative to Cloc, Tn ∈ so that X Cloc for all n. Then for each n there are (Tn,p)p 1 such that Tn,p ∈ Tn Tn,p ≥ ↑ ∞ a.s. as p and (X ) C for all p. For all n there is pn such that → ∞ ∈ 1 T < T n P[ n,pn n ] n ∧ ≤ 2 since a.s. convergence implies convergence in probability. Define a new sequence of stopping times S : T V T . Then, for all n, S S and n = n m n m,pm n n+1 ∧ ≥ ≤ X 1 1 S < T n ∞ . P[ n n ] 2m = 2n 1 ∧ ≤ m=n − By the Borel-Cantelli lemma, P[Sn < Tn n i.o.] = 0, which implies that Sn S T T S n n n,pn n ∧ ↑ ∞ a.s. Finally, X = ((X ) ) C since C is stable under stopping, so (Sn)n 1 is a localizing sequence for X relative∈ to C. ≥ 1.6.3 Corollary. (i) If X is a locally square integrable martingale then X is a local martingale (but the converse does not hold). (ii) If M is a process and there are stopping times (Tn)n 1 such that Tn a.s. and M Tn is a martingale for all n then M is a local martingale.≥ ↑ ∞ 1.7. Integration with respect to processes of finite variation 21 We will see that it is important to be able to decide when a local martingale a “true” martingale. To this end, let X t∗ := sups t Xs and X ∗ = sups 0 Xs . ≤ | | ≥ | | 1.6.4 Theorem. Let X be a local martingale. (i) If E[X t∗] < for all t then X is a martingale. ∞ (ii) If E[X ∗] < then X is a u.i. martingale. ∞ PROOF: Let (Tn)n 1 be a localizing sequence for X . For any s t, ≥ ≤ X X E[ Tn t s] = Tn t ∧ |F ∧ by the optional stopping theorem. Since X is dominated by the integrable ran- Tn t ∧ dom variable X t∗, we may apply the (conditional) dominated convergence theorem to both sides. Therefore X is a martingale. If E[X ∗] < then the family (X t )t 0 ∞ ≥ is dominated by the integrable random variable X ∗, so it is u.i. 1.6.5 Corollary. (i) If X is a bounded local martingale then it is a u.i. martingale. (ii) If X is a local martingale and a Lévy process then X is a martingale. (iii) If (X n)n 1 is a discrete time local martingale and E X n < for all n then X is a martingale.≥ | | ∞ n 1.6.6 Example. Suppose (An)n 1 is a measurable partition of Ω with P[An] = 2− ≥ for all n. Further suppose (Zn)n 1 is a sequence of random variables independent ≥n n of the An and such that P[Zn = 2 ] = P[Zn = 2− ] = 1/2. Define − ¨ σ(An : n 1) for 0 t < 1 t := ≥ ≤ F σ(An, Zn : n 1) for t 1 ≥ ≥ P completed with respect to . Define Y : Z 1 and T : 1S . P n = 1 k n k An n = 1 k n Ak ≤ ≤ ∞ ≤ ≤ Let X t be zero for t < 1 and Y for t 1. Then (Tn)n 1 is a localizing sequence Tn ∞ ≥ ≥ for X and X t is zero for t < 1 and Yn for t 1. Since Yn is bounded for each n, Tn ≥ 1 X is a u.i. martingale, but X is not a martingale because X1 = Y / L . ∞ ∈ 1.7 Integration with respect to processes of finite variation 1.7.1 Definition. We say that a process A is increasing or of finite variation if it has that property path-by-path for almost all ω. 1.7.2 Definition. Let A be of finite variation. The total variation process is 2n X A t : sup At k 1 2 n Atk2 n . = ( + ) − − | | n 1 k 1 | − | ≥ = Note that A t < a.s., A is increasing, and if A is adapted then so is A . | | ∞ | | | | 22 Preliminaries 1.7.3 Definition. Let A be of finite variation and F(ω, s) be ( )-measurable and bounded. F ⊗ B Z t (F A)t (ω) := F(s, ω)dAs(ω), · 0 the path-by-path Lebesgue-Stieltjes integral (which is well-defined and finite for almost all ω). The integral process is also of finite variation and if A is (right) continuous then the integral process is (right) continuous. If F is continuous path-by-path then then integral may be taken to be the Riemann-Stieltjes integral. 1.7.4 Theorem. Let A and C be adapted, increasing processes such that C A is increasing. There exists H jointly measurable and adapted such that A = H −C. · PROOF (SKETCH): The correct process is A(ω, t) A(ω, r t) H(t, ω) := lim . r 1,r − Q+ C(ω, t) C(ω, r t) ↑ ∈ − 1.7.5 Corollary. If A is of finite variation then there is a jointly measurable process H with 1 H 1 such that A = H A and A = H A. − ≤ ≤ · | | | | · ROOF KETCH + P (S ): Let A := ( A + A)/2 and A− := ( A A)/2. These are both | | | +| − + increasing processes, so by Theorem 1.7.4 there are H and H− such that A = + + H A and A− = H− A . Take H := H H−. · | | · | | − 1.7.6 Theorem. (i) Let A be of finite variation, H have continuous paths, and (Πn)n 1 be a sequence of partitions of [0, t] with mesh size converging to zero. For≥ each ti Πn let si denote any number in the interval [ti, ti+1]. Then ∈ Z t X lim Hs At At Hs dAs. n i ( i+1 i ) = →∞ Πn − 0 (ii) (Change of variable formula.) Let A be of finite variation with continuous 1 paths and f C (R). Then (f (At ))t 0 is of finite variation and ∈ ≥ Z t f (At ) = f (A0) + f 0(As)dAs. 0 (iii) Let A be of finite variation with continuous paths g be a continuous function. R t R At Then then g A dA g s ds. 0 ( s) s = 0 ( ) 1.8. Naïve stochastic integration is impossible 23 PROOF (OF (II)): Let (Πn)n 1 be a sequence of partitions of [0, t] as in (i) and consider ≥ X f A f A f A f A ( t ) ( 0) = ( ti+1 ) ( ti ) − Πn − X f A A A for some s by the MVT = 0( si )( ti+1 ti ) i Πn − Z t f 0(As)dAs by part (i) → 0 1.8 Naïve stochastic integration is impossible 1.8.1 Theorem (Banach-Steinhaus). Let X be a Banach space and let Y be a normed linear space. Suppose (Tα)α I is a collection of continuous linear opera- ∈ tors from X to Y such that supα I Tα x Y < for all x X . Then it is the case ∈ k k ∞ ∈ that supα I Tα L(X ,Y ) < . ∈ k k ∞ 1.8.2 Theorem. Let x : [0, 1] R be right continuous and let (Πn)n 1 be a se- quence of partitions of [0, 1] with→ mesh size converging to zero. If ≥ X S h : h t x x n( ) = ( i)( ti+1 ti ) Πn − converges for all h C([0, 1]) then x is of finite variation. ∈ PROOF: Note that Sn is a linear operator from C([0, 1]) to R, and X S x x . n ti+1 ti k k ≤ Πn | − | This upper bound is achieved for each n by the piecewise linear function hn with the property that h t sign x x . If S h exists for every h then the ( i) = ( ti 1 ti ) n( ) + − Banach-Steinhaus theorem implies that supn 1 Sn < , i.e. that the total varia- tion of x over [0, 1] is finite. ≥ k k ∞ Remark. It should be possible to prove that if X is a right continuous process and (Πn)n 1 is a sequence of partitions of [0, t] mesh size converging to zero P ≥ then if Ht X t X t converges in probability for all H with ti ,ti+1 Πn i ( i i 1 ) continuous paths∈ then X −is of− finite variation a.s. ∈ F ⊗ B Chapter 2 Semimartingales and stochastic integration 2.1 Introduction 2.1.1 Definition. A process H is a simple predictable process if it has the form n 1 X H H 1 t − H 1 t t = 0 0 ( ) + i (Ti ,Ti+1]( ) { } i=1 where 0 T T T < are stopping times and H L for all = 0 1 n i ∞( Ti ) i = 0, . . . , n. The≤ collection≤ · · · ≤ of all simple∞ predictable processes will∈ be denotedF S.A norm on S is H u := sups 0 Hs . When we endow S with the topology induced ≥ ∞ by u we writek k Su. k k k · k Remark. S . ⊆ P 0 0 2.1.2 Definition. L := L (Ω, , P) := X : X is finite-valued a.s. with the topology induced by convergenceF in probability{ ∈ F under P. } 0 0 Remark. L is not locally convex if P is non-atomic. The topological dual of L in this case is 0 . { } 2.1.3 Definition. Let H S have its canonical representation. For an arbitrary stochastic process X , the ∈stochastic integral of H with respect to X is n 1 X I H : H X − H X X . X ( ) = 0 0 + i( Ti+1 Ti ) i=1 − 0 A càdlàg adapted process X is called a total semimartingale if the map IX : Su L is continuous. X is a semimartingale if X t is a total semimartingale for all t →0. ≥ 24 2.2. Stability properties of semimartingales 25 2.2 Stability properties of semimartingales 2.2.1 Theorem. (i) semimartingales and total semimartingales are vector spaces. (ii) {If Q P then any} (total){ semimartingale under} P is also a (total) semi- martingale under Q. PROOF: (i) Sums and scalar multiples of continuous operations are continuous. (ii) Convergence in P-probability implies convergence in Q-probability. 2.2.2 Theorem (Stricker). Let X be a semimartingale with respect to the filtra- tion F and is a sub-filtration of F (i.e. t t for all t). If X is G adapted then X is a semimartingaleG with respect to G.G ⊆ F PROOF: S(G) S(F) and the restriction of a continuous operation is continuous. ⊆ 2.2.3 Theorem. Let A = (Aα)α I be a collection of pairwise disjoint events from ∈ and let t = σ( t A). If X is a semimartingale with respect to F then X is a semimartingaleF H withF respect∨ to H. Remark. This theorem is called Jacod’s countable expansion theorem in the text- book. PROOF: Note first that t = σ( t (A A A : P[A] = 0 ) since F is complete, and second that A AH : P[A]F ∨1/n \{is finite∈ because the} events are pairwise disjoint. It follows{ ∈ that we may≥ assume} without loss of generality that A is a S countable collection (An)n 1. If C := ∞n 1 An then we may assume without loss of C = generality that C A, i.e.≥ that A is a (countable) partition of Ω into events with positive probability.∈ For each n define Qn := P[ An], so that Qn P. By Theorem 2.2.1 X is an n ·| (F, Qn)-semimartingale. Let J be the filtration F completed with respect to Qn. n Claim. X is a (J , Qn)-semimartingale. n n Indeed, if H S(J , Qn) has its canonical representation then the Ti are J - ∈ n stopping times and the Hi are in L∞(Ω, T , Qn). J i n Fact. Any J -stopping time is Qn-a.s. equal to a F-stopping time. n Denote the corresponding F-stopping times by Ti0. Then any Hi T is Qn-a.s. ∈ J i equal to an r.v. in T (technical exercise). This shows that H is Qn-a.s. equal to a F i0 process in S(F, Qn), which proves the claim. n Note finally that Qn[Am] 0, 1 for all Am A, so F H J for all n. By ∈ { } ∈ ⊆ ⊆ Stricker’s theorem, X is an (H, Qn)-semimartingale for all n. It can be shown that, since d P A d , X is an , -semimartingale. P = ∞n=1 P[ n] Qn (H P) 26 Semimartingales and stochastic integration 2.2.4 Theorem. Let X be a process and (Tn)n 1 be a sequence of random times Tn ≥ such that X is a semimartingale for all n and Tn a.s. Then X is a semimartin- gale. ↑ ∞ PROOF: Fix t > 0. There is n0 such that P[Tn t] < " for all n n0. Suppose that k k k ≤ T ≥ H S are such that H 0. Let k0 be such that P[ IX n0 (H ) > α] < " for ∞ all k∈ k0. Then k k → | | ≥ k k k I t H > α I t H > α; T t I Tn t H > α; T > t < 2" P[ X ( ) ] = P[ X ( ) n0 ] + P[ (X 0 ) ( ) n0 ] | | | | ≤ | | k for all k k0. Therefore IX t (H ) 0 in probability, so X is a semimartingale. ≥ → 2.3 Elementary examples of semimartingales 2.3.1 Theorem. Every adapted càdlàg process with paths of finite (total) varia- tion is a (total) semimartingale. PROOF: Suppose that the total variation of X is finite. It is easy to see that, for all R H S, I H H ∞ d X . X ( ) 0 s ∈ | | ≤ k k∞ | | 2.3.2 Theorem. Every càdlàg L2-martingale is a total semimartingale. PROOF: Without loss of generality X0 = 0. Let H S have the canonical represen- tation. ∈ n 1 X 2 I H 2 − H X X E[( X ( )) ] = E i( Ti+1 Ti ) i=1 − n 1 X − H2 X X 2 X is a martingale = E i ( Ti+1 Ti ) i=1 − n 1 X H 2 − X X 2 u E ( Ti+1 Ti ) ≤ k k i=1 − H 2 X 2 X is a martingale = u E[ Tn ] k k2 2 H u E[X ] by Jensen’s inequality ≤ k k ∞ k k 2 Therefore if H H in Su then IX (H ) IX (H) in L , so also in probability. → → 2.3.3 Corollary. (i) A càdlàg locally square integrable martingale is a semimartingale. (ii) A càdlàg local martingale with bounded jumps is a semimartingale. (iii) A local martingale with continuous paths is a semimartingale. (iv) Brownian motion is a semimartingale. (v) If X = M + A is càdlàg, where M is a locally square integrable martingale and A is locally of finite variation, then X is a semimartingale. Such an X is called a decomposable process (vi) Any Lévy process is a semimartingale by the Lévy-Itô decomposition. 2.4. The stochastic integral as a process 27 2.4 The stochastic integral as a process 2.4.1 Definition. (i) D := adapted, càdlàg processes (ii) L := {adapted, càglàd processes} (iii) If C is{ a collection of processes then} bC is the collection of bounded processes from C. 2.4.2 Definition. For processes H n and H we say that H n H uniformly on com- pacts in probability (or u.c.p.) if → h n i lim P sup Hs Hs " = 0 n 0 s t | − | ≥ →∞ ≤ ≤ n (p) for all " > 0 and all t. Equivalently, if (H H)∗t 0 for all t. − −→ Remark. If we define a metric X 1 d X , Y : ∞ X Y 1 ( ) = 2n E[( )∗n ] n=1 − ∧ then u.c.p. convergence is compatible with this metric. We will denote by Ducp, Lucp, and Sucp the topological spaces D, L, and S endowed with the u.c.p. topology. The u.c.p. topology is weaker than the uniform topology. It is important to note that Ducp and Lucp are complete metric spaces. 2.4.3 Theorem. S is dense in Lucp. Rn PROOF: Let Y L and define Rn := inf t : Yt > n . Then Y Y u.c.p. R (exercise) and Y∈ n is bounded by n because{ Y is| left| continuous.} Thus →bL is dense in Lucp, so it suffices to show that S is dense in bLucp. Assume that Y bL and define Zt := limu t Yn for all t 0, so that Z D. For ∈ ↓ " ≥ ∈ each " > 0 define a sequence of stopping times T0 := 0 and " " Tn 1 := inf t > Tn : Zt ZT " > " . + { | − n | } " We have seen that the Tn are stopping times because they are hitting times for the " càdlàg process Z. Also because Z D, Tn a.s. (this would not necessarily happen it T " were defined in the same∈ way but↑ ∞ with Y ). Let " X Z : ZT " 1 T " ,T " . = n [ n n+1) n 0 ≥ Then Z" is bounded and Z" Z uniformly on compacts as " 0. Let → → " X U : Y01 0 ZT " 1 T " ,T " . = + n ( n n+1] { } n 0 ≥ 28 Semimartingales and stochastic integration " Then U Y01 0 + Z uniformly on compacts as " 0. If we define → { } − → n n," X Y : Y01 0 ZT " 1 T " ,T " = + k ( k k+1] { } k 1 ≥ n," then Y Y u.c.p. as " 0 and n . → → → ∞ 2.4.4 Definition. Let H S and X be a càdlàg process. The stochastic integral Pn T T process of H with respect∈ to X is J H : H X H X i+1 X i when X ( ) = 0 0 + i=1 i( ) Pn 1 − H H01 0 − Hi1 T ,T . = + i=1 ( i i+1] { } R R t Notation. We write H dX : H X : J H and H dX : I t H J H s s = = X ( ) 0 s s = X ( ) = ( X ( ))t R · and ∞ H dX : I H . 0 s s = X ( ) 2.4.5 Theorem. If X is a semimartingale then JX : Sucp Ducp is continuous. → PROOF: We begin by proving the weaker statement, that JX : Su Ducp is continu- ous. Suppose that H k 0. Let δ > 0 and define a sequence→ of stopping times k k ∞ k k k k → k S k T := inf t : (H X )t δ . Then H 1[0,T ] and H 1[0,T ] 0 because { | · | ≥ k} ∈ k k∞ → this already happens for (H )k 1. Let t 0 be given. ≥ ≥ k k k H X H X I t H k P[( )∗t δ] P[( )∗t T k δ] = P[ X ( 1[0,T ]) δ] 0 · ≥ ≤ · ∧ ≥ ≥ → k since X is a semimartingale. Therefore JX (H ) 0 u.c.p. Assume H k 0 u.c.p. Let " > 0, t > 0, and→δ > 0 be given. There is η such → k k that H < η implies P[JX (H)∗t δ] < ". Let R := inf s : Hs > η . Then k ∞ k k k k k ˜ k ≥ ˜ { | | } R a.s., H := H 1[0,R ] 0 u.c.p., and H η. → ∞ → k k ≤ k k k k k P[(H X )∗t δ] = P[(H X )∗t δ; R < t] + P[(H X )∗t δ; R t] · ≥ k · ≥ ˜ k · ≥ ≥ P[R < t] + P[(H X )∗t δ] < 2" ≤ · ≥ for k large enough. n n If H L and (H )n 1 S are such that H H u.c.p. then (Hn)n 1 has the ∈ ≥ ⊆ → n≥ Cauchy property for the u.c.p. metric. Since JX is continuous, (JX (H ))n 1 also ≥ has the Cauchy property. Since Ducp is complete there is JX (H) D such that n ∈ JX (H ) JX (H) u.c.p. We say that JX (H) is the stochastic integral of H with respect to→X . 2.4.6 Example. Let B be standard Brownian motion and let (Πn)n 1 be a se- quence of partitions of [0, t] with mesh size converging to zero. Define≥ X Bn : B 1 , = ti (ti ,ti+1] Πn 2.5. Properties of the stochastic integral 29 so that Bn B u.c.p. (check this as an exercise). → X J Bn B B B ( B( ))t = ti ( ti+1 ti ) Πn − 1 X 2 2 1 X 2 = (Bt Bt ) (Bt Bt ) 2 i+1 i 2 i+1 i Πn − − Πn − 1 2 1 X 2 = B (Bt Bt ) 2 t 2 i+1 i − Πn − (p) 1 2 1 Bt t −→ 2 − 2 R t 1 1 Hence B dB B t. 0 s s = 2 s 2 − 2.5 Properties of the stochastic integral 2.5.1 Theorem. Let H L and X be a semimartingale. ∈ T T (i) If T is a stopping time then (H X ) = (H1[0,T] X ) = H (X ). (ii) ∆(H X ) is indistinguishable from· H (∆X ). · · (iii) If · then H X is indistinguishable· from H X . Q P Q P (iv) If P λ · with λ 0 and P λ 1 then· H X is indistinguish- Q = k 1 k Pk k k 1 k = Q able from ≥H X for any k≥for which λ≥ > 0. · Pk k (v) If X is both a· P and Q semimartingale then there is H X that is a version of both H X and H X . · P Q (vi) If is another· filtration· and H then H X H X . G L(G) L(F) G = F (vii) If X has paths of finite variation∈ on compacts∩ then H ·X is indistinguishable· from the path-by-path Lebesgue-Stieltjes integral. · (viii) Y := H X is a semimartingale and K Y = (KH) X for all K L. · · · ∈ PROOF (OF (VIII)): It can be shown using limiting arguments that K (H X ) = (KH) X when H, K S. To prove that Y = H X is a semimartingale when· ·H L there· are two steps.∈ First we show that if K · S then K Y = (KH) X (and∈ this makes sense even without knowing that Y is∈ a semimartingale).· Fix· t > 0 and choose H n S with H n H u.c.p. We know H n X H X u.c.p. because X is a ∈ → · → · nk semimartingale. Therefore there is a subsequence such that (Y Y )∗t 0 a.s., nk nk − → nk where Y := H X . Because of the a.s. convergence, (K Y )t = limk (K Y )t n n n a.s. Finally, K Y k· = (KH k ) X since H , K S, so · →∞ · · · ∈ K Y u.c.p.-lim KH nk X KH X = k ( ) = ( ) · →∞ · · since KH nk KH u.c.p. and X is a semimartingale. For the second→ step, proving that Y is a semimartingale, assume that Gn G n → in Su. We must prove that (G Y )t (G Y )t in probability for all t. We have n · → · G H GH in Lucp, so → lim Gn Y lim GnH X GH X G Y. n = n ( ) = ( ) = →∞ · →∞ · · · Since this convergence is u.c.p. this proves the result. 30 Semimartingales and stochastic integration 2.5.2 Theorem. If X is a locally square integrable martingale and H L then H X is a locally square integrable martingale. ∈ · PROOF: It suffices to prove the result assuming X is a square integrable martingale, Tn Tn since if (Tn)n 1 is a localizing sequence for X then (H X ) = H X , so we would have that H ≥ X is a local locally square integrable· martingale.· Conclude with Theorem 1.6.2· , since the collection of locally square integrable martingale is stable under stopping. Furthermore, we may assume that H is bounded because left continuous processes are locally bounded (a localizing sequence is Rn := inf t : { Ht > n ) and that X0 = 0. | |Construct,} as in Theorem 2.4.3, H n S such that H n H u.c.p. By construc- tion the H n are bounded by H . Let t∈> 0 be given. It is→ easy to see that H n X is a martingale and k k∞ · n 2 X T n 2 n i+1 Ti E[(H X ) ] = E Hi X t X t · i=1 − 2 2 H E[X t ] < . ≤ k k∞ ∞ n (The detail are as in the proof of Theorem 2.3.2.) It follows that ((H X )t )n 1 is n · ≥ u.i. Since we already know (H X )t (H X )t in probability, the convergence is actually in L1. This allows us to· prove→ that· H X is a martingale, and by Fatou’s lemma · H X 2 lim inf H n X 2 H 2 X 2 < E[( )t ] n E[( )t ] E[ t ] · ≤ · ≤ k k∞ ∞ so H X is a square integrable martingale. · 2.5.3 Theorem. Let X be a semimartingale. Suppose that (Σn)n 1 is a sequence of random partitions, : 0 T n T n T n where the ≥T n are stopping Σn = 0 1 kn k times such that ≤ ≤ · · · ≤ (i) lim T n a.s. and n kn = (ii) sup →∞T n T∞n 0 a.s. k k+1 k | − | → P T n T n R Then if Y (or ) then Y n X k+1 X k Y dX . L D k Tk ( ) ∈ − → − Remark. The hard part of the proof of this theorem is that the approximations to Y do not necessarily converge u.c.p. to Y (if they did then the theorem would be a trivial consequence of the definition of semimartingale). 2.5.4 Example. Let Mt = Nt λt be a compensated Poisson process (a martin- gale) and let H 1 (a bounded− process in ), where T is the first jump time = [0,T1) D 1 R t of the Poisson process. Then H dM λ t T , which is not a local mar- 0 s s = ( 1) tingale. Examples of this kind are part of the− reason∧ that we want our integrands from L. 2.6. The quadratic variation of a semimartingale 31 2.6 The quadratic variation of a semimartingale 2.6.1 Definition. Let X and Y be semimartingales. The quadratic variation of X and Y is [X , Y ] = ([X , Y ]t )t 0, defined by ≥ Z Z [X , Y ] := XY X dY Y dX . − − − − Write [X ] := [X , X ]. The polarization identity is sometimes useful. 1 [X , Y ] = ([X + Y ] [X ] [Y ]) 2 − − 2.6.2 Theorem. 2 2 (i) [X ]0 = X0 and ∆[X ] = (∆X ) . (ii) If (Σn)n 1 are as in Theorem 2.5.3 then ≥ X n n u.c.p. 2 Tk 1 Tk 2 X0 + (X + X ) [X ]. k − −→ (iii) [X ] is càdlàg, adapted, and increasing. (iv) [X , Y ] is of finite variation, [X , Y ]0 = X0Y0, ∆[X , Y ] = ∆X ∆Y , and X T n T n T n T n u.c.p. X0Y0 + (X k+1 X k )(Y k+1 Y k ) [X , Y ]. k − − −→ (v) If T is any stopping time then T T T T T [X , Y ] = [X , Y ] = [X , Y ] = [X , Y ] . PROOF: R 0 (i) Recall that X : 0, so X dX 0 and X X 2. For any t > 0, 0 = 0 s s = [ ]0 = ( 0) − − 2 2 2 2 (∆X t ) = (X t X t ) = X t + X t 2X t X t − − 2 2− − − = X t X t 2X t ∆X t −2 − − − = (∆X )t 2∆(X X )t 2 − − · = ∆(X 2(X X ))t = ∆[X ]t − − · n 2 Rn 2 (ii) Fix n and let Rn := supk Tk . Then (X ) X u.c.p., so apply Theo- rem 2.5.3 and the summation-by-parts trick. → (iii) Let’s see why [X ] is increasing. Fix s < t rational. If we can prove that [X ]s [X ]t a.s. then we are done since [X ] is càdlàg. Use partitions (Σn)n 1 in Theorem≤ 2.5.3 that include s and t. Excluding terms from the sum makes≥ it smaller since all of the summands are squares (hence non-negative), so [X ]s [X ]t a.s. ≤ 32 Semimartingales and stochastic integration 2.6.3 Corollary. (i) If X is a continuous semimartingale of finite variation then [X ] is constant, 2 i.e. [X ]t = X0 for all t. (ii) [X , Y ] and XY are semimartingales, so semimartingales form an algebra. { } 2.6.4 Theorem (Kunita-Watanabe identity). Let X and Y be semimartingales and H, K : Ω [0, ) R be -measurable. × ∞ → F ⊗ B Z 2 Z Z ∞ ∞ 2 ∞ 2 Hs Ks d [X , Y ] s Hs d[X ]s Ks d[Y ]s. 0 | || | | | ≤ 0 0 PROOF: Use the following lemma. 2.6.5 Lemma. Let α, β, γ : [0, ) R be such that α(0) = β(0) = γ(0) = 0, α is of finite variation, β and γ are∞ increasing,→ and for all s t, ≤ 2 (α(t) α(s)) (β(t) β(s))(γ(t) γ(s)). − ≤ − − Then for all measurable functions f and g and all s t, ≤ Z t 2 Z t Z t 2 2 fu gu d α u fu dβu gu dγu. s | || | | | ≤ s s The lemma can be proved with the following version of the monotone class theo- rem. 2.6.6 Theorem (Monotone class theorem II). Let C be an algebra of bounded real valued functions. Let H be a collection of bounded real valued functions closed under monotone and uniform convergence. If C H then bσ(C) H. ⊆ ⊆ Both are left as exercises. 2.6.7 Definition. Let X be a semimartingale. The continuous part of quadratic c variation, [X ] , is defined by c 2 X 2 [X ]t = [X ]t + X0 + (∆Xs) . 0 PROOF: The Lebesgue-Stieltjes integration-by-parts formula gives us Z Z 2 X = X dX + X dX , − and by definition of [X ] we have Z 2 X = 2 X dX + [X ]. − Noting that Z Z Z X 2 X dX = (X + ∆X )dX = X dX + (∆X ) − − P 2 we obtain [X ] = (∆X ) , so X is a quadratic pure jump semimartingale. 2.6.10 Theorem. If X is a locally square integrable martingale that is not constant 2 everywhere then [X ] is not constant everywhere. Moreover, X [X ] is a local martingale and if [X ] 0 then X 0. − ≡ ≡ Remark. This theorem holds when X is a local martingale, but it is harder to prove. 2 R PROOF: We have that X [X ] = 2 X dX is a locally square integrable martin- − − 2 gale by Theorem 2.5.2. Assume that [X ]t = 0 for all t 0. Then X is a locally ≥ square integrable martingale, so suppose that (Tn)n 1 is a localizing sequence. Then X 2 X 2 0 for all n and all t. Thus X 2 ≥ 0 a.s. for all n and all t, E[ t Tn ] = 0 = t Tn = ∧ ∧ so X 0. If X0 is a constant other than zero then the proof applies to the locally ≡ square integrable martingale X X0. − 2.6.11 Corollary. Let X be a locally square integrable martingale and S T be stopping times. ≤ (i) If [X ] is constant on [S, T] [0, ) then X is constant there too. (ii) If X has continuous paths of∩ finite∞ variation on (S, T) then X is constant on [S, T] [0, ). ∩ ∞ T S PROOF: Consider M := X X and apply Theorems 2.6.2 and 2.6.10. − 2.6.12 Corollary. (i) If X and Y are locally square integrable martingales then [X , Y ] is the unique adapted, càdlàg process of finite variation such that a) XY [X , Y ] is a local martingale, and b) ∆[X−, Y ] = ∆X ∆Y and [X , Y ]0 = X0Y0. (ii) If X and Y are locally square integrable martingales with no common jumps and such that XY is a local martingale then [X , Y ] X0Y0. ≡ 34 Semimartingales and stochastic integration (iii) If X is a continuous square integrable martingale and Y is a square inte- grable martingale of finite variation then [X , Y ] X0Y0 and XY is martin- gale. ≡ PROOF: R R (i) By definition, XY [X , Y ] = X dY + Y dX , which is a local martingale. The second condition− is Theorem− 2.6.2(vi).− Conversely, if A satisfied the two conditions then, by Theorem 2.6.2(vi), A [X , Y ] is a continuous semi- − Fill in more of this proof. martingale of finite variation null at zero. By Corollary 2.2.3, A = [X , Y ]. 2.6.13 Assumption. [M] always exists for a local martingale M. 2.6.14 Corollary. Let M be a local martingale. Then M is a square integrable 2 martingale if and only if E[[M]t ] < for all t 0. We have E[Mt ] = E[[M]t ] ∞ ≥ when E[[M]t ] < . ∞ 2 PROOF: Suppose M is a square integrable martingale. M [M] is a local martin- − gale null at zero by Corollary 2.6.12. Let (Tn)n 1 be a localizing sequence. For all t 0, ≥ 2 2 ≥ M t lim M t T lim M M < E[[ ] ] = n E[[ ] n ] = n E[ t Tn ] E[ t ] ∧ →∞ →∞ ∧ ≤ ∞ The first equality is the monotone convergence theorem, the second is because 2 Tn t (M [M]) ∧ is a martingale, and the inequality is the “limsup” version of Fatou’s − lemma. For the converse, define stopping times Tn := inf t : Mt > n n. { | | } ∧ p sup M Tn n M n M s + ∆ Tn + [ ]n 0 s t | | ≤ | | ≤ ≤ ≤ Tn 2 1 Hence sups t Ms L L for all n and all t. By Doob’s maximal inequality. . . ≤ | | ∈ ⊆ 2.6.15 Example (Inverse Bessel process). Let B be a three dimensional Brown- ian motion that starts at x = 0. It can be shown that Mt := 1/ Bt is a local 6 2 k2 k martingale. Furthermore, E[Mt ] < for all t and limt E[Mt ] = 0. This implies that M is not a martingale (because∞ if it were then→∞M 2 would be a sub- 2 martingale, which contradicts that limt E[Mt ] = 0). Moreover, E[[M]t ] = →∞ 2 ∞ for all t > 0. It can be shown that M satisfies the SDE dMt = Mt dBt . − 2.6.16 Corollary. Let X be a continuous local martingale. Then X and [X ] have the same intervals of constancy a.s. (i.e. pathwise, cf. Corollary 2.6.11). 2.6.17 Theorem. Let X be a quadratic pure jump semimartingale. For any semi- martingale Y , X [X , Y ] = X0Y0 + ∆Xs∆Ys. 0 PROOF (SKETCH): We know by the Kunita-Watanabe inequality that d[X , Y ]s c c d[X , X ]s a.s. Since [X , X ] 0, this implies that [X , Y ] 0, so [X , Y ] is equal to the sum of its jumps. ≡ ≡ 2.6.18 Theorem. Let H, K L and X and Y be semimartingales. (i) [H X , K Y ] = (HK)∈[X , Y ]. · · · (ii) If H is càdlàg and (σn)n 1 is a sequence of random partitions tending to the identity then ≥ Z X T n T n T n T n u.c.p. H n X i+1 X i Y i+1 Y i H d X , Y . Ti ( )( ) s [ ]s − − −→ − Remark. Part (i) is important for computations, but part (ii) is important theoret- ically. We will prove Itô’s formula using part (ii). PROOF (OF (I)): It is enough to prove that [H X , Y ] = H [X , Y ] since we already have associativity. Suppose that H n S are· such that·H n H u.c.p. Define n n n Z := H X and Z := H X , so that Z ∈ Z u.c.p. Then → · · → Z Z n n n n [Z , Y ] = YZ Y dZ Z dY − − − − Z Z n n n = YZ (YH ) dX Z dY − − − − Z Z YZ (YH) dX Z dY → − − − − = [Z, Y ] For all n, since H n is simple predictable, n n n [Z , Y ] = [H X , Y ] = H [X , Y ] · n (check this). Finally, [X , Y ] is a semimartingale so H [X , Y ] H [X , Y ]. → · 2.7 Itô’s formula We have seen that if A is continuous and of finite variation and if f C 1 then ∈ Z t f (At ) = f (A0) + f 0(As)dAs. 0 We also saw that if B is a standard Brownian motion then Z t 2 Bt = 2 Bs dBs + t. 0 36 Semimartingales and stochastic integration 2 Take f (x) = x so that we can write Z t 1 Z t f (Bt ) f (B0) = f 0(Bs)dBs + f 00(Bs)d[B]s. − 0 2 0 This expression does not coincide with the one for continuous processes of finite variation. Notation. We write Z Z t Z t Z t Hs dXs := Hs dXs := Hs dXs H0X0 = Hs1(0, )(s)dXs. (0,t] 0+ 0 − 0 ∞ 2 2.7.1 Theorem (Itô’s formula). If X is a semimartingale and f C (R) then ∈ (f (X t ))t 0 is a semimartingale and ≥ Z t 1 Z t f X f X f X dX f X d X c ( t ) ( 0) = 0( s ) s + 2 00( s ) [ ]s − 0+ − 0+ − X + (f (Xs) f (Xs ) f 0(Xs )∆Xs). 0 Z t 1 Z t f X f X f X dX f X d X . ( t ) ( 0) = 0( s) s + 2 00( s) [ ]s − 0+ 0 Additionally, if X is of finite variation then Z t X f (X t ) f (X0) = f 0(Xs )dXs + (f (Xs) f (Xs ) f 0(Xs )∆Xs), − 0+ − 0 PROOF: First assume that X is bounded by a constant K. By definition the quadratic P 2 variation of a semimartingale is finite valued, so 0 Z t 1 Z t f X f X f X dX f X d X ( t ) ( 0) = 0( s ) s + 2 00( s ) [ ]s − 0+ − 0+ − X 2 + (f (Xs) f (Xs ) f 0(Xs )∆Xs f 00(Xs )(∆Xs) ) 0 2 where the remainder satisfies R(x, y) r( x y )(x y) , where r is an increas- | | ≤ | − | − ing function such that lim" 0 r(") = 0. P 2 ↓ Since 0 n n X 1 2 S S f X T n X T n X T n f X T n X T n X T n S + R = 0( i )( i+1 i ) + 00( i )( i+1 i ) n 2 Σ − − X 1 2 f 0(X T n )(X T n X T n ) + f 00(X T n )(X T n X T n ) i i+1 i 2 i i+1 i − Ln − − X R X T n , X T n . + ( i i+1 ) Sn Rn ∪ P 2 P 2 As n , n X T n X T n Xs . The remainder term in the above S ( i+1 i ) s B(∆ ) expression,→ ∞ call it Rn, satisfies− → ∈ n X R R X T n , X T n = ( i i+1 ) | | Sn Rn ∪ X 2 r X T n X T n X T n X T n ( i i+1 )( i+1 i ) ≤ Sn Rn | − | − ∪ X X 2 r X T n X T n X T n X T n = + ( i i+1 )( i+1 i ) Sn Rn | − | − X 2 X 2 r(2K) (X T n X T n ) + sup r( X T n X T n ) (X T n X T n ) i+1 i n i+1 i i+1 i ≤ Sn − R | − | Rn − 38 Semimartingales and stochastic integration nm By the properties of r we can pick a subsequence (Σ )m 1 such that ≥ For each m 1 there is nm such that ≥ X n 2 X 2 1 2 X T n X n Xs 1 " . ( i+1 Ti ) (∆ ) + n m σ m B=0 − − x B ≤ ∩ 6 ∈ You can find a sequence of partitions σ˜ m, constructed by refining σnm outside the subintervals that don’t contain jumps, such that X R X T˜ m , X T˜ m 0 ( i i+1 ) m σ˜ A=∅ → ∩ as m and " 0. Work from then on with the refinements σ˜ m. → ∞ → n 2 lim sup R m (1 + r(2K))" . m | | ≤ →∞ Hence nm nm nm f (X t ) f (X0) = SL + SS + SR − 1 nm X 2 S f X T n X T n X T n f X T n X T n X T n = L + 0( i )( i+1 i ) + 00( i )( i+1 i ) n 2 Σ m − − X 1 2 nm f 0(X T n )(X T n X T n ) + f 00(X T n )(X T n X T n ) + R i i+1 i 2 i i+1 i − Ln − − As m , → ∞ Z t X f X T n X T n X T n f Xs dXs by Theorem 2.5.3 0( i )( i+1 i ) 0( ) nm − Σ − → 0+ Z t X 1 2 1 f X T n X T n X T n f Xs d X s by Theorem 2.6.18 00( i )( i+1 i ) 00( ) [ ] nm 2 2 − Σ − → 0+ and similarly, n X X S f X T n f X T n f Xs f Xs L = ( i+1 ) ( i ) ( ) ( ) Ln − → s A − − X X∈ f X T n X T n X T n f Xs Xs 0( i )( i+1 i ) 0( )∆ Ln − → s A − ∈ X 1 2 X 1 2 f 00(X T n )(X T n X T n ) f 00(Xs )(∆Xs) 2 i i+1 i 2 Ln − → s A − ∈ Finally, as " 0, A(", t) exhausts the jumps of X . Since → 1 2 f (Xs) f (Xs ) f 0(Xs )∆Xs + f 00(Xs )(∆Xs) − − − − 2 − is dominated by a constant times the sum of the jumps squared. . . 2.7. Itô’s formula 39 k If X is not necessarily bounded then let Vk := inf t : X t > k and Y := X 1 . Then Y k is a semimartingale, as it is a product{ of| two| semimartingales,} [0,Vk) k Vk bounded by k. (Note that Y is not the same as X −, as the former jumps to zero at time Vk.) By the previous part, Z t 1 Z t f Y k f Y k f Y k dY k f Y k d Y k c ( t ) ( 0 ) = 0( s ) s + 2 00( s ) [ ]s − 0+ − 0+ − X k k k k + (f (Ys ) f (Ys ) f 0(Ys )∆Ys ). 0 2.7.2 Theorem. 1 n n (i) If X = (X ,..., X ) is an n-dimensional semimartingale and f : R R has 1 → n continuous second order partial derivatives then f (X ) = (f (X t ,..., X t ))t 0 is a semimartingale and ≥ n Z t n Z t 2 X ∂ f i 1 X ∂ f i j c f (X t ) f (X0) = (Xs )dXs + (Xs )d[X , X ]s ∂ xi − 2 ∂ xi∂ x j − − i=1 0+ i=1 0+ n X X ∂ f i + f (Xs) f (Xs ) (Xs )∆Xs − ∂ xi − 0 Z t 1 Z t f Z f Z f Z dZ f Z d Z c ( t ) ( 0) = 0( s ) s + 2 00( s ) [ ]s − 0+ − 0+ − X + (f (Zs) f (Zs ) f 0(Zs )∆Zs). 0 XY X0Y0 = X Y + Y X . − − ◦ − ◦ 40 Semimartingales and stochastic integration 2.8 Applications of Itô’s formula 2.8.1 Theorem (Stochastic exponential). Let X be a semimartingale with X0 = 0. Then there exists a unique semimartingale R t Z such that Z 1 Z dX . Moreover, t = + 0 s s − Y Z X : exp X 1 X c 1 X exp X . t = ( )t = ( t 2 [ ]t ) ( + ∆ s) ( ∆ s) E − s t − ≤ 1 c PROOF: Let’s see first that X is a semimartingale. We know that X X is ( ) 2 [ ] a semimartingale, so exp XE 1 X c is a semimartingale by Itô’s theorem.− We ( 2 [ ] ) Q − ∆Xs need to show that s t (1 + ∆Xs)e− is well-defined and is a semimartingale. ≤ 1 Towards this, let Ut Xs1 X 1 . The set s : Xs > is finite a.s., so it is = ∆ ∆ s 2 ∆ 2 Q | |≤ U { | | } 1 enough to prove that 1 U e s is a semimartingale. For x , it can be s t ( + s) − 2 ≤ 2 | | ≤ shown that log(1 + x) x x , whence | − | ≤ Y X Us log (1 + Us)e− log(1 + Us) Us s t ≤ s t | − | ≤ ≤ X 2 X 2 Us (∆Xs) [X ]t < a.s. ≤ s t ≤ s t ≤ ∞ ≤ ≤ Q ∆Xs Hence the infinite product converges, and moreover, s t (1 + ∆Xs)e− is of finite variation because the series of its log-jumps is absolutely≤ summable. In particular it is a quadratic pure jump semimartingale. x Now we prove that Z satisfied the integral equation. Let F(x, y) := ye , 1 c Q X 2 K : X X , and S : 1 X e ∆ s . Then F is C and Z F K , S . t = 2 [ ] t = s t ( +∆ s) − t = ( t t ) By Itô’s formula,− ≤ Z t Z t 1 Z Z 1 Z dK eKs dS Z d K c t = s s + − s + 2 s [ ]s − 0+ − 0+ 0+ − X Ks + (Zs Zs Zs (∆Ks) e − ∆Ss) 0 The only tricky point is to note that Zs = Zs (1 + ∆Xs). − Remark. (i) If X is of finite variation then so is (X ). E 2.8. Applications of Itô’s formula 41 (ii) If X is a local martingale then so is (X ), but at this point we can only prove it if X is a locally square integrableE martingale. (iii) Let T = inf t : ∆X t = 1 . { − } a) (X ) = 0 on [0, T) E 6 b) (X ) = 0 on [0, T] E − 6 c) (X ) = 0 on [T, ) (iv) If X isE continuous and∞X 0 then X exp X 1 X . 0 = ( ) = ( 2 [ ]) E − (v) If λ and B is standard Brownian motion then λB exp λB 1 λ2 t R ( ) = ( t 2 ) is a continuous∈ martingale. E − 2.8.2 Theorem (Yor’s formula). If X and Y are semimartingales with X0 = Y0 = 0 then (X ) (Y ) = (X + Y + [X , Y ]). E E E PROOF: Integration-by-parts. 1 Remark. If X is continuous then ( (X ))− = ( X + [X ]). E E − 1 n 2.8.3 Theorem. Let X = (X ,..., X ) be an n-dimensional local martingale with n i j values in D, an open subset of R . Assume [X , X ] = Aδi,j, where A is some increasing process. If u : D R is harmonic (resp. sub-harmonic) then u(X ) is a local martingale (resp. sub-martingale).→ 2.8.4 Theorem (Lévy’s characterization of BM). A stochastic process X is a Brownian motion if and only if X is a continuous local martingale with [X ]t = t a.s. for all t. PROOF: Only one direction requires proof, so assume X is a continuous local mar- 1 u2 t s iu(X t Xs) 2 ( ) tingale and [X ]t = t for all t. It suffices to show that E[e − s] = e− − for all 0 s < t < (exercise). Fix u , let F x, t : exp iux| F 1 u2 t , and let R ( ) = ( + 2 ) ≤ ∞ ∈ Zt := F(X t , t). By Itô’s formula, Z t u2 Z t u2 Z t Zt Z0 = iu Zs dXs + Zs ds Zs d[X ]s − 0 2 0 − 2 0 Z t Zt = 1 + iu Zs dXs 0 Therefore Z is a local martingale by Theorem 2.5.2. But 1 2 1 u2 t sup Z sup exp iuX u t e 2 < . t = ( t + 2 ) = s t | | s t | | ∞ ≤ ≤ so Z is a martingale, and this implies what we wanted to prove. 42 Semimartingales and stochastic integration 2.8.5 Theorem (Multidimensional Lévy’s characterization). 1 n If X = (X ,..., X ) is a vector of n continuous local martingales and, for all i, j, i j [X , X ]t = tδi,j for all t then X is an n-dimensional Brownian motion. 2.8.6 Theorem. Let M be a continuous local martingale with M0 = 0 and such that limt [M]t = . Let →∞ ∞ (i) Ts = inf t : [M]t > s , (ii) : { and } , and s = Ts G = ( s)g 0 (iii) GB : MF . G ≥ s = Ts Then B is a -Brownian motion, M is -stopping time for all t, and M B G [ ]t G t = [M]t a.s. for all t. PROOF: Since [M] = , Ts < for all s and B is well-defined. Also, since T T for s u,∞ ∞ ∞ , so really is a filtration. It is right s u Ts = s Tu = u G continuous≤ because≤ FT is rightG continuous.⊆ F G For all s and t, M s T t , [ ]t = s Ts = s { ≤ } { ≥ } ∈ F G so the M are -stopping times. M Ts M s for all s since M is [ ]t G E[[ ] ] = E[[ ]Ts ] = [ ] a continuous process. By Corollary 1.4.8, M∞ Ts is a u.i. -martingale with M 2 F E[ Ts ] = M s. By the optional sampling theorem E[[ ]Ts ] = B M M B E[ s u] = E[ Ts Tu ] = Tu = u |G |F T 2 T for all u s, so B is a G-martingale. One can prove that (M s ) [M] s is a u.i. martingale≤ for each s (exercise). Hence − B2 B2 M 2 M 2 M M s u. E[ s u u] = E[ T T Tu ] = E[[ ]Ts [ ]Tu Tu ] = − |G s − u |G − |G − 2 This implies that (Bu u)u 0 is a G-martingale. By Corollary 2.6.16 M and [M] − ≥ have the same intervals of constancy, so B = MT is a continuous process. Corol- lary 2.6.12 implies that [B]u = u, so B is a Brownian motion by Lévy’s charac- terization. (See the exercises around p.99 in the text regarding time changes.) Finally, note that T t because T is the right continuous inverse of M , and [M]t [ ] T > t if and only if ≥M is constant on the interval t, T , which happens if [M]t [ ] ( [M]t ) and only if M is constant on that interval. Hence B M for all t. [M]t = t 2.8.7 Exercise. If B is a G-Brownian motion then B is also a G˜ -Brownian motion, ˜ T T where s = u>s u. Note that E[Z u>s u] = limu s E[Z u] (this equality is knownG as Lévy’s theoremG ). | G ↓ |G 2.8.8 Theorem (Lévy’s stochastic area formula). Let B = (X , Y ) be a standard R t R t 2-dimensional Brownian motion and let A : X dY Y dX . Then eiuAt t = 0 s s 0 s s E[ ] = 1/ cosh(ut) for all u and 0 t < . − ≤ ∞ Chapter 3 The Bichteler-Dellacherie Theorem and its connexions to arbitrage 3.1 Introduction Classical semimartingales 3.1.1 Definition. An adapted càdlàg process S is a classical semimartingale if it can be decomposed as S = S0 +M +A, where M0 = A0 = 0, M is a local martingale, and A is of finite variation. Remark. Being a classical semimartingale is a local property, i.e. if S is a process for which there is a sequence of stopping times (Tn)n 1 such that Tn a.s. and STn is a classical semimartingale for all n then S is a classical≥ semimartingale.→ ∞ Say T n n P n S n M A . If one defines M : M 1 then one can prove that M = + = n 1 (Tn 1,Tn] is a local martingale and A : S M P≥ An1 − is of finite variation. = = n 1 (Tn 1,Tn] − ≥ − 3.1.2 Theorem (Bichteler ’79, Dellacherie ’80). For an adapted, càdlàg, real val- ued process S the following are equivalent. (i) S is a semimartingale. (ii) S is a classical semimartingale. The following, seemingly stronger, theorem is true. This is the theorem that is stated in the textbook and the one that we will eventually prove. 3.1.3 Theorem. For an adapted, càdlàg, real valued process S the following are equivalent. (i) S is a semimartingale. (ii) S is decomposable (i.e. S = M + A where M is a locally square integrable martingale and A is of finite variation.) (iii) For all β > 0 there are M and A such that M is a local martingale with jumps bounded by β and A is of finite variation and S = M + A. (iv) S is a classical semimartingale. 43 44 The Bichteler-Dellacherie Theorem and its connexions to arbitrage So far we have seen that (ii) implies (i) (cf. Corollary 2.3.3) and (iii) implies (ii) (because a local martingale with bounded jumps is locally bounded, and in particular is a locally square integrable martingale). The theoretically important, and difficult, results are (iv) implies (iii) and (i) implies (iv). The former is impor- tant enough to be restated as a theorem. 3.1.4 Theorem (Fundamental theorem of local martingales). If M is a local martingale then for all β > 0 there are N and B such that M = N +B, where N is a local martingale with jumps bounded by β and B is of finite variation. Connexion to finance Suppose that S is a price process, i.e. St is the price of an asset at time t, where we make no assumptions about S other than that it is càdlàg. Let Ht be the number of units of S that you hold at time t, so that H is a simple predictable process. The accumulated gains and losses after following the strategy H up to R t time t are H S H dS (the integral makes sense for any S when H is a ( )t = 0 s s simple predictable· process). An arbitrage over the time horizon [0, T] is a strategy H such that H0 = 0, (H S)T 0 a.s., and P[(H S)T > 0] > 0. It is very important to be able to determine· when≥ there are no possible· arbitrages. 3.1.5 Definition. We say that S allows a free lunch with vanishing risk, or FLVR, n for simple integrands if there is a sequence of simple predictable processes (H )n 1 such that both of the following conditions hold. ≥ n + (FL) (H S)T does not converge to zero in probability. · n (VR) limn (H S)− u = 0. Alternatively,→∞ Sk satisfies· kno free lunch with vanishing risk, or NFLVR, for simple inte- n grands if, for all sequences of simple predictable processes (H )n 1, (VR) implies n + ≥ (H S)T 0 in probability. · → It is clear that if S admits an arbitrage then S allows FLVR. In 1994 Delbaen and Schachermayer proved that NFLVR for simple integrands implies S is a semi- martingale. Their proof relies on the Bichteler-Dellacherie theorem. 3.1.6 Definition. We say that S allows free lunch with vanishing risk for simple integrands with little investment, or FLVR+LI if there is a sequence of simple pre- n dictable processes (H )n 1 such that (FL) and (VR) hold and so does n ≥ (LI) limn H u = 0. Alternatively,→∞ kS satisfiesk no free lunch with vanishing risk and little investment, or NFLVR+LI, for simple intgrands if, for all sequences of simple predictable processes n n + (H )n 1, (VR) and (LI) together imply (H S)T 0 in probability. ≥ · → 3.1.7 Theorem. For a locally bounded, adapted, càdlàg process S the following are equivalent. (i) S satisfies NFLVR+LI. (ii) S is a classical semimartingale. 3.2. Proofs of Theorems 3.1.7 and 3.1.8 45 3.1.8 Theorem. For an adapted càdlàg process S the following are equivalent. n (i) For all sequences (H )n 1 of simple predictable processes, n ≥ a) limn H u = 0, i.e. (LI) →∞ k k n b) limn sup0 t T (H S)−t = 0 in probability (this is slightly weaker than→∞ (VR)). ≤ ≤ · n + together imply (H S)T 0 in probability. (ii) S is a classical semimartingale.· → Remark. The assertions (ii) implies (i) in both Theorems 3.1.7 and 3.1.8 follow from (iv) implies (i) in Theorem 3.1.3, because (i) in each theorem holds triv- ially for semimartingales. Also note that (i) implies (iv) in Theorem 3.1.3 is a consequence of Theorem 3.1.8. 3.2 Proofs of Theorems 3.1.7 and 3.1.8 First we want to show that if S is locally bounded, càdlàg, and adapted and satis- fies NFLVR+LI then S is a classical semimartingale. Note that the collection of processes satisfying NFLVR+LI is stable under stop- ping. Since the property of being a classical semimartingale is local and S is locally bounded, we may assume that S u 1. Without loss of generality we assume k k ≤ that S0 = 0 and the time interval is [0, 1]. Notation. Let Dn denote the dyadic rationals in [0, 1] with denominator at most n n n S 2 , i.e. Dn := j2− : j = 0, . . . , 2 . Let D := n 1 Dn denote the collection of dyadic rationals{ in [0, 1]. } ≥ n n n n n For all n 1 define S := M + A , where M and A are defined on the set ≥ n n n Dn inductively by A0 := 0, M0 = S0, and for j = 1, . . . , 2 , n n A n A n : Sj2 n S j 1 2 n j 1 2 n (3.1) j2 (j 1)2 = E[ − ( ) − ( ) − ] − − − − − − n −n |F Mj2 n := Sj2 n Aj2 n . (3.2) − − − − Then M n is a martingale and An is predictable and of finite variation. The hope is that we may pass to the limit as n in some sense. We are assuming that S is locally bounded, so there are two cases.→ ∞ n n CASE 1. If (M )n 1 and (A )n 1 remain bounded in some sense then, by using Komlós’ lemma,≥ one can≥ pass to the limit to obtain a local martingale M and a predictable process A of finite variation such that S = M + A. n n CASE 2. If (M )n 1 or (A )n 1 does not remain bounded then neither does the other, and one≥ can construct≥ a sequence of strategies witnessing that S al- lows FLVR+LI. The idea is that if a predictable process can “get close to” a martingale then this leads to an arbitrage. 3.2.1 Theorem. Let (fn)n 1 be a sequence of random variables. ≥ 1 (KOMLÓS’ LEMMA). If (fn)n 1 is bounded in L , i.e. supn 1 fn 1 < , then there is a subsequence f ≥ such that 1 f f ≥ convergesk k ∞ a.s. as k . ( nk )k 1 k ( n1 + + nk ) ≥ ··· → ∞ 46 The Bichteler-Dellacherie Theorem and its connexions to arbitrage 2 (MAZUR’SLEMMA). If (fn)n 1 is bounded in L , i.e. supn 1 fn 2 < , then there ≥ ≥ k k ∞ 2 exist gn conv(fn, fn+1,... ) such that (gn)n 1 converges a.s. and in L . ∈ ≥ PROOF (OF MAZUR’SLEMMA): Let H be the Hilbert space generated by the fn’s. For 2 all n define Kn = conv(fn, fn+1,... ), where the closure is taken in the L -norm. In fact, Kn coincides with the weak closure of conv(fn, fn+1,... ) by a theorem from functional analysis. By the Banach-Alaoglu theorem we know that Kn is weakly T compact for all n. Since Kn+1 Kn for all n, n 1 Kn = ∅ by the finite intersection ⊆ ≥ 6 property of compact sets. Hence there is a sequence of gn conv(fn, fn+1,... ) 2 T ∈ such that (gn)n 1 converges in L to an element of n 1 Kn. By passing to a ≥ ≥ subsequence we may assume that (gn)n 1 converges a.s. ≥ Remark. We can apply Mazur’s lemma to countably many sequences simultane- m 2 ously. More precisely, suppose that (fn )n 1 is bounded in L for all m. Then for all n there are convex weights λn,..., λ≥ n , independent of m, such that the ( n Nn ) sequences λn f m λn f m converge a.s. and in L2 for all m. ( n n + + Nn n )n 1 ··· ≥ 3.2.2 Proposition. Let S = (St )0 t 1 be càdlàg and adapted, with S0 = 0 and ≤ ≤ n n such that S u 1 and S satisfies NFLVR+LI. Let A and M be as defined in (3.1) k k ≤ and (3.2). For all " > 0 there is C > 0 and a sequence of stopping times (ρn)n 1 such that, for all n, ≥ (i) ρn takes values in Dn . (ii) ρn < < ". ∪ {∞} P[ ] n,ρ ∞ An,ρn M n,ρn n M n 2 C (iii) The stopped processes and satisfy, for all , 1 L2 and k k ≤ 2n X n,ρn n,ρn n,ρn TV(A ) := Aj2 n A j 1 2 n C. − ( ) − j=1 − − ≤ 3.2.3 Lemma. Under the same assumptions as in Proposition 3.2.2, with 2n n X 2 Q : Sj2 n S j 1 2 n , = ( − ( ) − ) − j=1 − n the sequence (Q )n 1 is bounded in probability (i.e. for all α > 0 there is M such n that P[Q M] < α≥ for all n). ≥ n n P2 PROOF: For all n, let H : S j 1 2 n 1 j 1 2 n,j2n , a simple predictable pro- = j=1 ( ) − (( ) − ] cess. Recall that a b a − 1 a b−2 1 a2− b2 . ( ) = 2 ( ) + 2 ( ) − − − − 2n n X H S t St j 1 2 n St j2 n St j 1 2 n ( ) = ( ) − ( − ( ) − ) ∧ − ∧ ∧ − · − j=1 − 2n 1 X 2 1 2 2 St j2 n St j 1 2 n S S = ( − ( ) − ) + ( 0 t ) 2 ∧ ∧ − 2 j=1 − − 3.2. Proofs of Theorems 3.1.7 and 3.1.8 47 1 n 1 1 2 = Q + (S0 St ) 2 2 − Clearly H n 1, and H n S 1 for all t because S 1 and Qn is u ( )t 2 u k k ≤ · ≥ − n k k ≤ 0 nonnegative. Assume for contradiction that (Q )n 1 is not bounded in L . Then ≥ nm there is α > 0 such that for all m > 0 there is nm such that P[(H S)1 m] α. nm · ≥ ≥ Whence (H /m)m 1 would be a FLVR+LI. ≥ For c > 0 define a sequence of stopping times k k X 2 σn c : inf : Sj2 n S j 1 2 n c 4 . ( ) = n ( − ( ) − ) 2 − j=1 − ≥ − Given " > 0 there is c1 such that P[σn(c1) < ] < "/2 by Lemma 3.2.3. ∞ 3.2.4 Lemma. Under the same assumptions as Proposition 3.2.2, the stopped n,σ c M n,σn(c1) M n( 1) 2 c martingales satisfy 1 L2 1. k k ≤ n n n PROOF: For n 1 and k = 1, . . . , 2 , since the A s are predictable and the M s are martingales, ≥ σn(c1) σn(c1) 2 σn(c1) σn(c1) 2 σn(c1) σn(c1) 2 E[(Sk2 n S k 1 2 n ) ] = E[(Mk2 n M k 1 2 n ) ] + E[(Ak2 n A k 1 2 n ) ] − ( ) − − ( ) − − ( ) − − − − − − − σn(c1) 2 σn(c1) 2 E[(Mk2 n ) (M k 1 2 n ) ] − ( ) − ≥ − − σn(c1) 2 Write E[(M1 ) ] as a telescoping series and simplify to get σn(c1) 2 X σn(c1) σn(c1) 2 2 M S S S S n E[( 1 ) ] = E[( k2 n k 1 2 n ) ] + E[( σn c1 σn c1 2 ) ] − ( ) − ( ) ( ) − k2 n c − − − − − σn( 1) ≤ 2 (c1 4) + 2 = c1. ≤ − 2n σ c 1 n n,σn(c1) P ( n( 1) ) n n 3.2.5 Lemma. Let V := TV(A ) = i 1 ∧ Aj2 n A j 1 2 n . Under the = − ( ) − n | − − | assumptions of Proposition 3.2.2 the sequence (V )n 1 is bounded in probability. ≥ n PROOF: Assume for contradiction that (V )n 1 is not bounded in probability. Then ≥ nk there is α > 0 such that for all k there is nk such that P[V k] α. For n 1 define ≥ ≥ ≥ n n,σn(c1) n,σn(c1) n bj 1 := sign Aj2 n A j 1 2 n j 1 2 − ( ) − ( ) − − − − ∈ F − n n P2 n n and H t : b 1 j 1 2n,j2 n t . Then H u 1 and ( ) = j=1 j 1 (( ) − ]( ) − − k k ≤ X σ c σ c σ c σ c n,σn(c1) n n( 1) n( 1) n n( 1) n( 1) (H S)t = bj2 n Sj2 n S j 1 2 n + b t2n St S t2n − − ( ) − · j t2n − − b c − b c ≤b c n,σn(c1) n n,σn(c1) n H A t2n 2 n H M t2n 2 n 2 ( ) − + ( ) − ≥ · b c · b c − 48 The Bichteler-Dellacherie Theorem and its connexions to arbitrage n,σn(c1) n n,σn(c1) n and at time t = 1 we have (H S)1 = V + (H M )1. But the second 2 summand is bounded in L (it is at most· c1 by Lemma 3.2.4·), so we conclude that n,σn(c1) (H S)1 is not bounded in probability. Define· a sequence of stopping times j n,σn(c1) n ηn(c) := inf n : (H M )j2 n c . 2 | · − ≥ n,σn(c1) n 2 Because sup n h M j2 n 4c1 by Doob’s sub-martingale in- E[( 1 j 2 (( ) − ) ] n,σn(c1≤) ≤ n · ≤ equality, (H M ) is bounded in probability. Therefore there is c0 > 0 such n, c c · σn( 1) ηn( 0) that P[ηn(c0) < ] α/2. Note that H ∧ S is (uniformly) bounded n, c c ∞ ≤ σn( 1) ηn( 0) · below by c0. We claim (H ∧ S)1 is not bounded in probability. Indeed, for any n and any k, · n,σn(c1) ηn(c0) α P[(H ∧ S)1 k] ≤ · ≥ n,σn(c1) P[(H S)1 k, ηn(c0) = ] + P[ηn(c0) < ]. ≤ · ≥ ∞ ∞ Since P[ηn(c0) < ] α/2, the probability of the other event is at least α/2. This gives the desired∞ contradiction≤ because it is now easy to construct a FLVR+LI. PROOF (OF PROPOSITION 3.2.2): Define a sequence of stopping times k k X n n τn(c) := inf n : Aj2 n A j 1 2 n c . 2 − ( ) − j=1 | − − | ≥ By Lemma 3.2.5 there is c2 such that P[τn(c2) < ] < "/2. Take C := c1 c2 and ∞ ∨ ρn := σn(c1) τn(c2). ∧ 3.2.6 Lemma. Let f , g : [0, 1] R be measurable functions, where f is left continuous and takes finitely many→ values. Say f PK f s 1 . Define = k=1 ( k) (sk 1,sk] − K X (f g)(t) := f (sk 1)(g(sk) g(sk 1)) + f (sk(t))(g(t) g(sk(t))) − − · k=1 − − where k(t) is the biggest of the k such that sk less than or equal to t. Then for all partitions 0 t0 tM 1, ≤ ≤ · · · ≤ ≤ M M X X (f g)(ti) (f g)(ti 1) 2 TV(f ) g + g(ti) g(ti 1) f . − ∞ − ∞ i=1 | · − · | ≤ k k i=1 | − | k k 3.2.7 Proposition. Let S = (St )0 t 1 be càdlàg and adapted, with S0 = 0 and such ≤ ≤ that S u 1 and S satisfies NFLVR+LI. For all " > 0 there is C and a [0, 1] k k ≤ n ∪{∞} valued stopping time α such that P[α < ] < " and sequences (M )n 1 and n ∞ ≥ (A )n 1 of continuous time càdlàg processes such that, for all n, ≥ n n (i) A0 = M0 = 0 3.2. Proofs of Theorems 3.1.7 and 3.1.8 49 α n,α n,α (ii) S = A + M n,α n,α 2 C (iii) M is a martingale with M1 L2 P2n n,α n,α k k ≤ (iv) j 1 Aj2 n A j 1 2 n C. = − ( ) − | − − | ≤ n n PROOF: Let " > 0 be given. Let C, M , A , and ρn be as in Proposition 3.2.2. n n n n n Extend M and A to all t [0, 1] by defining Mt := E[M1 t ] and At = St Mt . Note that the extended An∈is no longer necessarily predictable,|F and currently− we n,ρn only have control of the total variation of A over Dn, i.e. n 2 (ρn 1) X∧ n n Aj2 n A j 1 2 n C. − ( ) − j=1 | − − | ≤ n n Notice that, for t ((j 1)2− , j2− ], ∈ − n n At = St Mt − n = St E[Mj2 n t ] − − |F n = St E[Sj2 n Aj2 n t ] − − n − − |F = Aj2 n (E[Sj2 n t ] St ) − − − |F − n n n,ρn From this and S u 1 it follows that At Aj2 n 2, so A u C + 2. k k ≤ k − − k∞ ≤ k k ≤ How do we find the “limit” of sequence of stopping times (ρn)n 1? The trick is to define Rn : 1 , a simple predictable process, and note that≥ stopping at = [0,ρn 1] n∧ n,ρn n n n,ρn n n ρn is like integrating R , i.e. A = R A and M = R M . We have that · · 1 Rn 1 1 ρ < 1 ". E[ 1] = E[ ρn= ] = P[ n ] ≥ ∞ − ∞ ≥ − Apply Komlós’ lemma to Rn to obtain convex weights µn,..., µn such that ( 1)n 1 ( n Nn ) ≥ n X∞ n i R := µi R1 R1 a.s. as n i=n → → ∞ By the dominated convergence theorem, E[R1] 1 ". Observe that ≥ − n X∞ n i i X∞ n i i R S = µi (R M ) + µi (R A ) · i=n · i=n · | {z } | {z } 2 L norm pC TV over Dn is C ≤ ≤ Define α : inf t : Rn 1 . Each Rn is a left continuous, decreasing process. n = t 2 In particular, Rn{ 1 >≤ 0,} so we can divide by this quantity. We claim that αn 2 α < < ". Indeed,≥ on the event α < , Rn Rn 1 so P[ n ] [ n ] 1 αn 2 ∞ ∞ ≤ + ≤ n n 1 " 1 R 1 R 1α < αn < . E[ 1] E[( αn+) n ] P[ ] ≥ − ≥ − ∞ ≥ 2 ∞ 50 The Bichteler-Dellacherie Theorem and its connexions to arbitrage Define new processes T n : 1 t /Rn. Then T n 2 and T n Rn S Sαn . t = [0,αn]( ) t u ( ) = Thus we define Mn and An by k k ≤ · · X X αn n ∞ n i i n ∞ n i i n n S = T µi (R M ) + T µi (R A ) =: M + A . · i=n · · i=n · n The total variation of T over Dn is bounded by 3. By Lemma 3.2.6, 2n X n n n X∞ n i i Aj2 n A j 1 2 n 2 TVn(T ) µi (R A ) − ( ) − j=1 | − − | ≤ i=n · ∞ n X∞ n i i + T TVn µi (R A ) ∞ k k i=n · 6(C + 2) + 2C ≤ n 2 C M n,αn 2 C That M1 L2 follows from the fact that 1 L2 . To finish the proof, we showk k that≤ there is a subsequence α k suchk that≤ α : inf α satisfies ( nk )k 1 = k nk α < 4". We know R 2 3" because≥ R 1 ". Since Rn R P[ ] P[ 1 3 ] E[ 1] 1 1 a.s. there∞ is≤ a subsequence such≤ that ≤ Rn R 1 "≥2 k−. Finally, → P[ 1 1 15 ] − | − | ≥ ≤ 3 nk P[α < ] P inf R1 ∞ ≤ k ≤ 5 3 2 nk 3" + P inf R1 , R1 > ≤ k ≤ 5 3 X 3 2 3" ∞ Rnk , R > + P 1 5 1 3 ≤ k=1 ≤ X 1 3" ∞ Rnk R + P 1 1 15 ≤ k=1 | − | ≥ 4" ≤ n n Therefore (M )n 1, (A )n 1, and α have the desired properties. ≥ ≥ Remark. One thing to take away from this is that if you need to take a “limit of stopping times” then one way to do it is to turn the stopping times into processes and take the limits of the processes. PROOF (OF (I) IMPLIES (II) IN THEOREM 3.1.7): We may assume the hypothesis of Proposition 3.2.7. Let " > 0 and take C, α, n n (M )n 1, and (A )n 1 as in Proposition 3.2.7. Apply Komlós’ lemma to find convex weights≥ λn,..., λn≥ such that ( n Nn ) n n,α n Nn,α λnM1 + + λN M1 M1 ··· n → n n,α n Nn,α λnAt + + λN At At ··· n → 3.2. Proofs of Theorems 3.1.7 and 3.1.8 51 for all t D, where the convergence is a.s. (and also in L2 for M). For all n, ∈ 2n X n,α n,α Aj2 n A j 1 2 n C − ( ) − j=1 | − − | ≤ α so the total variation of A over D is bounded by C. Further, we have S = Mt + At (where Mt = E[M1 t ]). A is càdlàg on D, so define it on all of [0, 1] to make it 2 càdlàg. M is an L martingale|F so it has a càdlàg modification. Since P[α < ] < " and " > 0 was arbitrary, and the class of classical semimartingales is local,∞S must be a classical semimartingale. Remark. On the homework there is an example of a (not locally bounded) adapted càdlàg process that satisfies NFLVR and is not a classical semimartingale. PROOF (OF (I) IMPLIES (II) IN THEOREM 3.1.8): We no longer assume that S is locally bounded. The trick is to leverage the re- sult for locally bounded processes by subtracting the “big” jumps from S. Assume without loss of generality that S 0 and define J : P S 1 . Then 0 = t = s t ∆ s ∆Ss 1 X := S J is an adapted, càdlàg, locally bounded process. We≤ will show| |≥ that The- orem 3.1.8− (i) for S implies NFLVR+LI for X , so that we may apply Theorem 3.1.7 to X . Then since J is of finite variation, this will then imply S is a classical semi- martingale. n n n Suppose H S are such that H u 0 and (H X )− u 0. We need to prove n ∈ k k → k · k →n that (H X )T 0 in probability. First we will show that (H S)− u 0. · → k · k → n n n sup (H S)−t sup (H X )−t + sup (H J)t 0 t T · ≤ 0 t T · 0 t T | · | ≤ ≤ ≤ ≤ n ≤ ≤n sup (H X )−t + ( H TV(J))T ≤ 0 t T · k k∞ · ≤ ≤ 0 by the assumptions on H n → n n By (i), (H S)T 0 in probability. Since (H J) 0 in probability (as J is · → n · →n n a semimartingale), we conclude that (H X )T = (H S)T (H J)T 0 in probability. Therefore X satisfies NFLVR+LI.· · − · → Remark. n n (i) S satisfies NFLVR if (H S)− u 0 implies (H S)T 0 in probability. This k · k → · → is equivalent to the statement that (H S)T : H S, H0 = 0, (H S)− u 1 is bounded in probability. The latter{ is sometimes· ∈ called no unboundedk · k profit≤ } with bounded risk or NUPBR. (ii) It is important to note that this definition of NFLVR does not imply that there is no arbitrage. The usual definition of NFLVR in the literature is as follows. Let K := (H S)T : H S, H0 = 0 and let C := f L∞ : f g for some { · ∈ } { ∈ ≤ g K . It is said that there is NFLVR if and only if C L∞ = 0 (the closure ∈ } ∩ + { } is taken in the norm topology of L∞). This is equivalent to NUPBR+NA, (where NA is no arbitrage, the condition that K L∞ = 0 ). ∩ + { } 52 The Bichteler-Dellacherie Theorem and its connexions to arbitrage 3.3 A short proof of the Doob-Meyer theorem 3.3.1 Definition. An adapted process S is said to be of class D if the collection of random variables Sτ : τ is a finite valued stopping time is uniformly integrable. { } 3.3.2 Theorem (Doob-Meyer for sub-martingales). Let S = (St )0 t T be a sub- ≤ ≤ càdlàg + predictable implies what? martingale of class D. Then S can be written uniquely as S = S0 + M + A where M Note that *-martingale means càdlàg. is a u.i. martingale with M0 = 0 and A is a càdlàg, increasing, predictable process with A0 = 0. We will require the following Komlós-type lemma to obtain a limit in the proof of the theorem. 3.3.3 Lemma. If (fn)n 1 is a u.i. sequence of random variables then there are ≥ 1 gn conv(fn, fn+1,... ) such that (gn)n 1 converges in L . ∈ ≥ k k 2 PROOF: Define f f 1 for all n, k . Then f L for all n and k since n = n fn k N n they are bounded random| |≤ variables. As∈ noted before,∈ there are convex weights λn,..., λn and f k L2 such that λn f k λn f k f k L2. Since f ( n Nn ) n n + + Nn Nn ( n)n 1 The details need to be filled in here. is u.i., ∈ ··· → ∈ ≥ lim sup λn f k λn f k λn f λn f 0 ( n n + + N N ) ( n n + + N Nn ) 1 = k n n n n →∞ k ··· − ··· k which implies that λn f λn f is Cauchy in L1. ( n n + + Nn Nn )n 0 ··· ≥ PROOF (OFTHE DOOB-MEYERTHEOREMFORSUB-MARTINGALES): Assume without loss of generality that T = 1. Subtracting the u.i. martingale (E[S1 t ])0 t 1 from S we may further assume that S1 = 0 and St 0 for all t. |Fn ≤ ≤ ≤ Define A0 = 0 and, for t Dn, ∈ n n At At 2 n := E[St St 2 n t 2 n ] − − − − − n −n − |F − Mt := St At . − An An is predictable and it is increasing since S is a sub-martingale. Fur- = ( t )t Dn ∈ n n n n thermore, M1 = A1 since S1 = 0, and in fact Mt = E[A1 t ] for all t Dn. Therefore for every− Dn valued stopping time τ, − |F ∈ n n Sτ = E[A1 τ] + Aτ. − |F n n We would like to show that (A1)n 1 is u.i. (and hence that (M1 )n 1 is u.i.). For ≥n n ≥ c > 0 define τn(c) := inf (j 1)2− : Aj2 n > c 1, which is a stopping time since − An is predictable. Then A{ − c and An > c } ∧ τ c < 1 , so τn(c) 1 = n( ) ≤ { } { } n n A 1 n A 1 E[ 1 A1>c] = E[ 1 τn(c)<1] An 1 = E[E[ 1 τn(c)] τn(c)<1] |F 3.3. A short proof of the Doob-Meyer theorem 53 S An 1 = E[( τn(c) + τ c ) τn(c)<1] − n( ) S 1 c τ c < 1 E[ τn(c) τn(c)<1] + P[ n( ) ] ≤ − Also, τn(c) < 1 τn(c/2) < 1 , so { } ⊆ { } n n Sτ c/2 1τ c/2 <1 A A 1τ c/2 <1 E[ n( ) n( ) ] = E[( 1 τn(c/2)) n( ) ] − n − n A A 1τ c <1 E[( 1 τn(c/2)) n( ) ] ≥ c − P[τn(c) < 1] ≥ 2 n Whence A 1 n 2 S 1 S 1 We may now ap- E[ 1 A >c] E[ τn(c/2) τn(c/2)<1] E[ τn(c) τn(c)<1] 1 ≤ − −n ply the class D assumption on S to show that (A1)n 1 is u.i., provided that we can ≥ show P[τn(c) < 1] 0 as c uniformly in n. But, for any n, → → ∞ n 1 n 1 n 1 P[τn(c) < 1] = P[A1 > c] E[A1] = E[M1 ] = E[S0] ≤ c − c − c n n which goes to zero as c . Therefore (A1)n 1, and hence also (M1 )n 1, are u.i. sequences of random variables.→ ∞ ≥ ≥ n n n Extend M to all t [0, 1] by defining Mt := E[M1 t ]. By Lemma 3.3.3 there are convex weights∈λn,..., λn and M L1 such that|F the sequence of martingales n Nn Mn : λn M n λn M Nn satisfies M∈n M in L1. Then Mn M : M Fill in the details (use Jensen’s in- = n + + Nn 1 t E[ t ] = t as well. ··· → → |F equality?). n P n n Extend A to on 0, 1 by defining it to be A 1 t 2 n,t . Note that A [ ] t Dn t ( − ] ∈ n − n n n Nn extended in this way is a predictable process. Define A := λnA + + λN A ··· n and At := St Mt . For all t D, − ∈ n n 1 At := (St Mt ) At in L − → and, by passing to a subsequence, we may assume the convergence is a.s. This im- plies in particular that (At )t D is increasing, so by right continuity, so is (At )t [0,1]. n ∈ n ∈ Further, A is predictable for all n. If we can prove that lim supn At (ω) = At (ω) a.s. for all t [0, 1] then this would imply that A is a predictable process and we would be done.∈ n n By right continuity lim supn At At for all t and limn At = At if A is contin- uous at t. Since A is càdlàg there are≤ countably many jump→∞ times. To prove what n we want it suffices to prove, for all stopping times τ, lim supn Aτ = Aτ a.s. Fix a n n 1 stopping time τ. Since Aτ A1 A1 in L , we may apply Fatou’s Lemma to get ≤ → n n n lim inf E[Aτ] lim sup E[Aτ] E[lim sup Aτ] E[Aτ] n ≤ n ≤ n ≤ n Therefore it is actually enough to show that limn E[Aτ] = E[Aτ]. Let σn := →∞ inf t Dn : t τ , a stopping time. Note that σn τ as n . By the definition of A{ n,∈An An≥, so} ↓ → ∞ τ = σn lim An lim An n E[ τ] = n E[ σn ] →∞ →∞ 54 The Bichteler-Dellacherie Theorem and its connexions to arbitrage lim Sσ Mσ discrete Doob-Meyer = n E[ n ] E[ n ] →∞ − = E[Sτ] E[M0] since S is class D − = E[Aτ]. 3.4 Fundamental theorem of local martingales 3.4.1 Definition. A stopping time T is predictable there there is a sequence of stopping times (Tn)n 1, called an announcing sequence, such that Tn < T on the ≥ event T > 0 and Tn T. A stopping time T is said to be totally inaccessible if P[T ={S < ]} = 0 for all↑ predictable stopping times S. ∞ 3.4.2 Example. Suppose that X is a continuous process. T := inf t : X t K is predictable and T : inf t : X K 1 is an announcing sequence.{ The≥ first} n = t n jump time of a Poisson process{ is≥ totally− inaccessible.} 3.4.3 Theorem. If S is a càdlàg sub-martingale of class D and S jumps only at totally inaccessible stopping times then S can be written uniquely as S = S0+M+A, where M is a u.i. martingale with M0 = 0, and A is continuous, increasing, and A0 = 0. PROOF: This is Theorem 10 in Chapter III of the textbook. The proof of uniqueness in this case is easy, since if M + A = S = M 0 + A0 are two decompositions then A A0 is a continuous martingale of finite variation null at zero. Hence A A0 0 by− Theorem 2.6.10. See the textbook for the proof of existence. − ≡ Remark. A stronger statement is true. If S is as in the statement of the Theo- rem 3.4.3 and S jumps at a totally inaccessible stopping time T then A in the decomposition is continuous at T. 3.4.4 Definition. Let A be an adapted process of finite variation with A0 = 0. If A satisfies any of the following equivalent conditions then A is said to be a natural process. (i) E[ A ] < (i.e. A is of integrable variation) and E[[M, A] ] = 0 for all bounded| |∞ martingales∞ M. ∞ R (ii) ∞ M dA M A for all all bounded martingales M. (This is an E[ 0 s s] = E[ ] integration-by-parts− formula,∞ ∞ in some sense.) R R (iii) ∞ M dA ∞ M dA for all all bounded martingales M. E[ 0 s s] = E[ 0 s s] R − (iv) ∞ M dA 0 for all all bounded martingales M. E[ 0 ∆ s s] = (v) [M, A] is a− martingale for all bounded martingales M. Remark. A càdlàg process locally of integrable variation is of finite variation. 3.4. Fundamental theorem of local martingales 55 3.4.5 Theorem. If A is a natural process and a martingale then A 0. ≡ PROOF: Fix a finite valued stopping time T and let H be a bounded, non-negative martingale. It can be shown that, since A is a martingale, Z T E Hs dAs = 0 0 − using dominated convergence and the fact that it is trivial if A is a simple pro- cess (take the limit of the Riemann sums). This implies, since A is natural, that H A 0. Now take H : 1 , a bounded, non-negative martingale. E[ T T ] = t = E[ AT >0 t ] Then |F 1 A 1 A H A 0. E[ AT >0 T ] = E[E[ AT >0 T ] T ] = E[ T T ] = |F Therefore AT 0. But since E[AT ] = 0, it must be the case that AT = 0. Since this holds for any≤ finite valued stopping time T, A 0. ≡ 3.4.6 Theorem. If A is a predictable process of integrable variation null at zero then A is a natural process. PROOF: This is Theorem 14 in the textbook. There a different proof in Jacod and Shiryaev as well. 3.4.7 Lemma. If Y is a u.i. martingale of finite variation then Y is locally of inte- grable variation. PROOF: Define τn := inf t : Yt > n . For any t we have Yt Y0 + Y t and hence { | | } | −| ≤ | | | | − (∆ Y )t = ∆Yt Yt + Yt Yt + Y0 + Y t . | | | | ≤ | | | −| ≤ | | | | | | − Thus Y Y Y 2 Y Y Y 2n Y Y L1. τn τn + (∆ )τn τn + 0 + τn + 0 + τn | | ≤ | | − | | ≤ | | − | | | | ≤ | | | | ∈ Therefore (τn)n 1 is a localizing sequence for Y relative to the class increasing processes integrable≥ at . | | ∞ 3.4.8 Theorem. If M is a predictable local martingale of finite variation then M is constant. n Tn PROOF: Suppose that (Tn)n 1 is a localizing sequence for M. If X := (M M0) n n n≥ − then X , X0 = 0, X is u.i., and it is of finite variation. By Lemma 3.4.7, X is locally∈ P of integrable variation. Without loss of generality assume that it is of integrable variation. By Theorem 3.4.6, X n is natural. By Theorem 3.4.5, X n is the zero process. 3.4.9 Corollary. The Doob-Meyer decomposition is unique. 56 The Bichteler-Dellacherie Theorem and its connexions to arbitrage 3.4.10 Theorem (Doob-Meyer without class D). Suppose that S is a super-martingale. Then S can be decomposed uniquely as S = S0 + M A where M is a local martingale with M0 = 0 and A is càdlàg, − increasing, and predictable with A0 = 0. m PROOF: Define T := inf t : St m m. By the optional sampling theorem, 1 T m 1 T m ST m L and S is dominated{ | | ≥ by m} ∧ ST m L . Hence S is of class D, T m m m m so by∈ the Doob-Meyer decomposition we∨ may| | write ∈ S = M A where M m m − is u.i. and A is increasing and predictable with A0 = 0. By the uniqueness in m m 1 m m 1 Doob-Meyer for class D, it must be the case that M = M + and A = A + on 0, T m . The desired processes may be defined as M : P M m1 and [ ] = ∞m=1 [Tm 1,Tm) A : P Am1 . − = ∞m=1 [Tm 1,Tm) − 3.5 Quasimartingales, compensators, and the fundamental theorem of local martingales 3.5.1 Definition. Let X be a process and be a partition for which X L1 Π ti ∈ for all ti Π. Define the variation of X along Π by var (X ) := E[C(X , Π)], P Π where by C∈ X , : X t X t t . We say that X is a quasimartingale if ( Π) = Π i E[ i+1 i ] var X : sup var X <| −(and hence|F | X < for all t 0, ). ( ) = Π Π( ) E[ t ] [ ) ∞ | | ∞ ∈ ∞ 3.5.2 Theorem (Rao). X on [0, ) is a quasimartingale if and only if X = Y Z where Y and Z are nonnegative super-martingales.∞ − PROOF: See the textbook, though the proof therein is not so clear. 3.5.3 Theorem. A quasimartingale X has a unique decomposition as X = M + A where M is a local martingale and A is a predictable process locally of integrable variation. PROOF: Existence and uniqueness follow from Theorems 3.4.10 and 3.5.2. Let’s Y Y see that A is locally of integrable variation. Suppose X = Y Z where Y = M A Z Z Y Z and Z = M A by the Doob-Meyer decomposition. Clearly− A = A A , so− we Y Z need only show− that A and A are locally of integrable variation. Let− τm be a localizing sequence for M Y . Y,τm Y,τm A lim At E[ ] = t E[ ] ∞ →∞ Y,τm τm = lim E[Mt ] E[Yt ] t − →∞ τm = Y0 lim E[Yt ] − t →∞1 Y0 L ≤ ∈ Z and the same argument works for A . 3.5. Quasimartingales, compensators, and the fundamental theorem of local martingales 57 3.5.4 Corollary (Existence of the compensator). Let A be a process locally of integrable variation. There is a unique finite variation process Ap such that A Ap is a local martingale. ∈ P − p p Remark. A is called the compensator of A. When A = [X , Y ] then A =: X , Y , the predictable variation between X and Y . 〈 〉 PROOF: A is locally a quasimartingale because it is locally of finite variation. 3.5.5 Examples. p (i) If N is a Poisson of process of rate λ then Nt = λt. We know what N [N] is a local martingale, but in this case [N] = N. What is meaningful is− that ([N]t λt)t 0 is a local martingale. − ≥ (ii) Let At := 1t τ = 1[τ, )(t), where τ is a random time. Let F be the minimal filtration that≥ satisfies∞ the usual conditions and makes τ a stopping time. (F 0 may be realized as follows. Let t := σ( τ s : s t) and take t τ T 0 F { ≤ }p R≤ ∨dF N(u) : .) The -compensator of A is A ∧ , where t = ">0 t " F t = 0 1 F u F F + ( ) F is the cumulative distribution function of τ on [0, ]. Note− that− if F is continuous then so is Ap. ∞ 3.5.6 Theorem (Existence of the predictable projection). Let X be a bounded process that is measurable. There exists a unique process pX such that (i) pX F ⊗ B p ∈ P (ii) ( X )T = E[X T T ] on T < for all predictable stopping times T. |F − { ∞} 3.5.7 Theorem. p p (i) If X is a local martingale then X = X and (∆X ) = 0. p p (ii) If A is locally of integrable variation and− càdlàg then (∆A) = ∆A . p PROOF (OF (II)): By the definition of the compensator, A A is a local martingale. p p p p p p By (i), (∆(A A )) = 0, so (∆A) = (∆A ). Since A −is itself predictable, so is p p p p ∆A , and hence− (∆A ) = ∆A . The following proof is based on the presentation in Jacod and Shiryaev. See the textbook for a different presentation. PROOF (OFTHEFUNDAMENTALTHEOREMOFLOCALMARTINGALES): Assume without loss of generality that M0 = 0. By localization we may assume that M is a u.i. martingale (this requires work, but we have done it previously). Let b β/2 and define A : P M 1 . The process A is well-defined = t = s t ∆ s ∆Ms b ≤ | |≥ and of finite variation since M is càdlàg. Let Tn := inf t : A t > n or Mt > n so that A A A n A and A { M| | n |M | . Hence} Tn = Tn + ∆ Tn + ∆ Tn ∆ Tn ∆ Tn + Tn A | |2n | M| − |L1 because| ≤ M| is u.i.,| so therefore| | ≤ |A is locally| ≤ integrable| | and Tn + Tn p p A| | exists.≤ Let| B :|= ∈A A and N := M B. Clearly B| is| of finite variation, and it is a local martingale− by definition of −Ap, and N is a local martingale because 58 The Bichteler-Dellacherie Theorem and its connexions to arbitrage it is a difference of local martingales. Define X : M 1 , so that A t = ∆ t ∆Mt 3.5.8 Corollary. (i) A classical semimartingale is a semimartingale. (ii) A càdlàg sub- (super-) martingale is a semimartingale. PROOF: Cf. Theorems 3.1.4, 2.3.1, 2.3.2 and the Doob-Meyer decomposition. 3.5.9 Theorem. If H L and M is a local martingale then H M is well-defined and is a local martingale.∈ · Remark. This theorem is not true if H / L. (Example?) This is an important difference between the general theory of∈ stochastic integration and the theory of stochastic integration with respect to continuous local martingales. In the latter the integral of any predictable process is a local martingale. PROOF: Write M = N + B as in the fundamental theorem of local martingales. By localization we may assume that H is bounded, M is u.i. and has bounded jumps, and that B has integrable variation (cf. lemma in the proof of Theorem 3.4.8). We know by Theorem 3.4.8 that H N is a locally square integrable martingale. Using the dominated convergence theorem· one can prove that H B is a martingale (use “Riemann sums”). · 3.6 Special semimartingales and another decomposition theorem for local martingales 3.6.1 Definition. A semimartingale X is said to be a special semimartingale if it can be written X = X0 + M + A where M0 = A0 = 0, M is a local martingale, and A is predictable and of finite variation. 3.6.2 Examples. (i) Quasimartingales (including sub- and super-martingales). (ii) Local martingales (iii) Continuous semimartingales (proved below). 3.6.3 Theorem. If X is a special semimartingale then the decomposition in the definition is unique. This decomposition is known as the canonical decomposition of X . 3.6.4 Theorem. The following are equivalent for a semimartingale X . (i) X is a special semimartingale. 3.6. Special semimartingales and another decomposition theorem for local martingales 59 (ii) X ∗, the running maximum of X , is locally integrable. (iii) J := (∆X )∗ is locally integrable. (iv) Decomposing X as X = M +A, where M is a local martingale and A has finite variation, implies that A is locally of integrable variation. (v) There exists a local martingale M and A locally of integrable variation such that X = M + A. PROOF: (I) IMPLIES (II) Recall the following fact. If A is càdlàg, predictable, and of finite variation then A is locally of integrable variation. (This is an exercise on the homework.) Recall also that if M is a local martingale then Mt∗ = sups t Ms ≤ | | is locally integrable. (Indeed, we may assume that M is u.i. Let Tn := inf t : M > n and note that M Tn n M , which is integrable.) If X is{ a t∗ ( ∗) Tn special semimartingale} then write≤ X∨= | M +| A as in the definition and note that X ∗ M ∗ + A∗ M ∗ + A , so X t∗ is locally integrable for all t. II IMPLIES III≤ ≤ | | ( ) ( ) For all s t, ∆Xs = Xs Xs 2X t∗, which implies J 2X ∗. (III) IMPLIES (IV) Homework.≤ | | | − −| ≤ ≤ (IV) IMPLIES (V) Trivial. (V) IMPLIES (I) Write X = M + A where A is locally of integrable variation. Then p p p by Corollary 3.5.4, A exists and X = (M + A A ) + A . Note that the first summand is a local martingale and the second,− Ap, is predictable and of finite variation. 3.6.5 Theorem. If X is a continuous semimartingale then X may be written X = M + A, where M is a local martingale and A is continuous and of finite variation. In particular, X is a special semimartingale. PROOF: By Theorem 3.6.4(iii) X is special, so write X = M + A in the unique way. p p Let’s see that A is continuous. We have A , so A = A and (∆A) = ∆A. But ∈ P p p p ∆A = (∆A) = (∆X ) (∆M) p − p = (∆X ) (∆M) = 0 by Theorem 3.5.7 = 0 X is continuous by assumption. 3.6.6 Definition. Two local martingales M and N are orthogonal if MN is a lo- cal martingale. In this case we write M N. A local martingale M is purely ⊥ discontinuous if M0 = 0 and M N for all continuous local martingales N. ⊥ ˜ Remark. If N is a Poisson process of rate λ then Nt := Nt λt is purely discontin- uous. However, N˜ is not the sum of its jumps, i.e. not locally− constant pathwise. 3.6.7 Theorem. Let H2 be the space of L2 martingales. 2 (i) For all M H , E[[M] ] < and M exists. ∈ ∞ ∞ 〈 〉 60 The Bichteler-Dellacherie Theorem and its connexions to arbitrage 2 (ii) (M, N)H2 := E[ M, N ] + E[M0N0] defines an inner product on H , and H2 is a Hilbert space〈 with〉∞ this inner product. T (iii) M N if and only if M, N 0, which happens if and only if M N N0 for⊥ all stopping times〈T. 〉 ≡ ⊥ − (iv) Take H2,c to be the collection of martingales in H2 with continuous paths and H2,d to be the collection of purely discontinuous martingales in H2. 2,c 2 2,c 2,d 2 2,c Then H is closed in H and (H )⊥ = H . In particular, H = H H2,d . ⊕ 3.6.8 Theorem. Any local martingale M has a unique decomposition M = M0 + c d c d M +M , where M is a continuous local martingale, M is a purely discontinuous c d local martingale, and M0 = M0 = 0. PROOF (SKETCH): Uniqueness follows from the fact that a continuous local martin- gale orthogonal to itself must be constant. For existence, write M = M0 + M 0 + M 00 as in the fundamental theorem of local martingales, where M 0 has bounded jumps and M 00 is of finite variation. Then M 00 is purely discontinuous and M 0 is locally 2 c c in H . But then M 0 = M + N where M is a continuous local martingale and 2,d d N H is purely discontinuous. Hence we may take M = M 00 + N. ∈ Suppose that h : R R and h(x) = x in a neighbourhood ( b, b) of zero, e.g. → − h(x) = x1 x <1. Let X be a semimartingale and notice that ∆Xs h(∆Xs) = 0 only | | − 6 if ∆Xs b. Define | | ≥ ˇ X X (h)t := ∆Xs h(∆Xs). 0 3.6.9 Definition. We call (B, C, ν) the semimartingale characteristics of X associ- ated with h, where (i) B := B(h) c c (ii) C := M(h) = [M(h) ], cf. Theorem 3.6.8, and note that C doesn’t depend on h by〈 the uniqueness〉 part of that theorem. (iii) ν is the “compensator” of the random measure µX associated with the jumps of X . X µ d t, d x : 1 δ d t, d x X ( ) = ∆Xs=0 (s,∆Xs)( ) s t 6 ≤ (Inspired by the Lévy-Khintchine formula and Lévy measures.) c c c Remark. M(h) and C do not depend on the function h. We write X := M(h) for the “continuous part” of the semimartingale X . (This does not mean that X X c is locally constant.) − 3.7. Girsanov’s theorem 61 3.6.10 Example. If X is the Lévy process that appears in the Lévy-Khintchine for- 2 2 mula and h(x) = x1 x <1 then B(h) = αt (α depends on h), C(h) = σ t (σ does not depend on h), and| | ν(d t, d x) = ν(d x) d t, where ν is the Lévy measure (which depends on h). 3.7 Girsanov’s theorem Working on the usual setup (Ω, , F = ( t )t 0, P), we know that if Q P and X is a P local martingale then XFis a Q semimartingaleF ≥ (cf. Theorem 2.2.1 ). By the Bichteler-Dellacherie theorem X = M + A = N + B where A and B have finite variation, M is a P local martingale, and N is a Q local martingale. The question is whether, knowing M and A, we can describe N and B. 3.7.1 Theorem. If for all t (we write in this case) and M is Q t P t Q loc P an adapted càdlàg process|F then|F the following hold. d Q t (i) There exists a unique Z 0 such that Zt = |F for all t. d P t ≥ |F (ii) MZ is a P martingale if and only if M is a Q martingale. (iii) If MZ is a P local martingale then M is a Q local martingale. (iv) If M is a Q local martingale with localizing sequence (Tn)n 1 such that ≥ P[limn Tn = ] = 1 then MZ is a P local martingale. →∞ ∞ Remark. A property being “P-local” subtly depends on the measure P via the re- quirement that the localizing sequence of stopping times (Tn)n 1 satisfies Tn ≥ ↑ ∞ P-a.s. If Q loc P then a property holds P-locally if and only if it holds Q-locally. ∼ 3.7.2 Corollary. If Q loc P then MZ is a P local martingale if and only if M is a Q local martingale. ∼ 3.7.3 Theorem (Girsanov-Meyer). Suppose that Q loc P and let Z be as in The- R 1 orem 3.7.1. If M is a local martingale then M ∼d Z, M is a local martin- P Z [ ] Q gale. − PROOF: By integration-by-parts, ZM [Z, M] = Z M +M Z. By Theorem 3.5.9, ZM Z, M is a local martingale.− By Corollary− 3.7.2· M− · 1 Z, M is a local [ ] P Z [ ] Q martingale.− By integration-by-parts, − 1 1 1 Z, M Z, M Z, M 1 , Z, M [ ] = [ ] + [ ] + [ Z [ ]] Z Z · − · Z − By Theorem 3.5.9 the middle term is a Q local martingale. We also have 1 1 X 1 1 Z, M 1 , Z, M Z, M Z, M Z, M [ ] + [ Z [ ]] = [ ] + ∆ ∆[ ] = [ ] Z · Z · Z Z · − − 62 The Bichteler-Dellacherie Theorem and its connexions to arbitrage where this last integral is a Lebesgue-Stieltjes integral, since [Z, M] has finite variation. Thus 1 Z, M is a local martingale plus 1 Z, M . Finally, Z [ ] Q Z [ ] · 1 1 1 1 M Z, M M Z, M Z, M Z, M Z [ ] = Z [ ] + Z [ ] Z [ ] − · | − {z } | −{z · } Q local mtg Q local mtg Remark. M M 1 Z, M 1 Z, M is a decomposition showing that M = ( Z [ ]) + Z [ ] is a Q-semimartingale.− · If M is a special· semimartingale then this might not be the canonical decomposition because it may not be the case that 1 Z, M is Z [ ] predictable. · 3.7.4 Theorem (Girsanov’s theorem, predictable version). Suppose that Q loc P and let Z be as in Theorem 3.7.1. If M is a P local martin- gale and [M, Z]is P locally of integrable variation and M, Z is the P-compensator of M, Z then M M 1 Z, M is -a.s. well-defined〈 and〉 is a local martin- [ ] 0 = Z Q Q c − ·〈 〉 c gale. Moreover, [M ] is a version of [(M 0) ]. 1 1 PROOF: Let T : inf t : Z < . Then A Z, M is well-defined on 0, T . n = t n = Z [ n] { } − 〈 〉 Let T = limn Tn. →∞ \ Q[T < ] Q[ Tn < ] ∞ ≤ n { ∞} lim Tn < = n Q[ ] →∞ ∞ P lim ZT 1T < = n E [ n n ] ∞ →∞ 1 lim = 0 n n ≤ →∞ (Below all of the arguments are on [0, Tn].) By integration-by-parts MZ = M Z + Z M + [Z, M] − · − · = M Z + Z M + [Z, M] Z, M + Z, M | − · {z − · } | {z− 〈 }〉 〈 〉 P local mtg (Thm 3.5.9) P local mtg Hence ZM Z, M is a P-local martingale. Now, A is predictable and of finite variation. − 〈 〉 AZ = A Z + Z A + [Z, A] = A Z + Z A − · − · · − · (Note that A Z is well-defined and, by results we will see in Chapter 4, a P local martingale since· A is locally bounded). But Z A = Z, M , so AZ Z, M is a P local martingale. Hence ZM ZA = ZM −Z·, M 〈 (ZA〉 Z, M −) is 〈 a P〉local − − 〈 〉 − − 〈 〉 martingale. By Theorem 3.7.1 M 0 = M A is a Q local martingale. c − c To see that [M ] is a version of [(M 0) ] on uses the fact that, for a local mar- c c tingale N, [N ] = [N] , which is to be proved as an exercise. (Assume that d P 2 [N ] = (∆N) when N0 = 0.) 3.7. Girsanov’s theorem 63 Remark. If Q loc P and X is a P-semimartingale then X is a Q-semimartingale. Indeed, if X = M + A is a decomposition under P then M = M 0 + M 00 where M 0 has bounded jumps and M 00 is of finite variation, so X is M 0 plus a finite variation process. You can prove that [Z, M 0] is locally of integrable variation. By the previous theorem, X is M 1 Z, M plus a finite variation process, so is a 0 Z 0 − − 〈 〉 Q-semimartingale. 3.7.5 Theorem. Suppose that and let Z be the density process d Q . Let Q loc P d P R := inf t : Zt = 0, Zt = 0 , X be a P-local martingale, and Ut := ∆XR1t R. Then R t 1 − p ≥ X { d X , Z U is} a -local martingale. t 0 Z [ ]s + t Q − s How to use Girsanov’s Theorem Suppose that X is a P-semimartingale and X = X0 + M + A is a decomposition with A = J M where J L (there may or may not be such a decomposition). Define d Q ·〈 〉 J M ∈Z, where dZ JZ dM. If the density process is a d P = ( ) = Q-martingaleE then− · − − − Z 1 M d Z, M is a Q-local martingale − Z 〈 〉 − d[Z, M] = JZ d[M] − − d Z, M = JZ d M 〈 Z 〉 − − 〈 〉 M + Jd M = X X0 is a Q-local martingale 〈 〉 − 3.7.6 Theorem. Let M be a local martingale (KAZAMAKI’SCRITERION). Let S be the collection of bounded stopping times. If M is continuous and sup exp 1 M < E[ ( 2 T )] T S ∞ ∈ then (M) is a u.i. martingale. 1 (NOVIKOV’SCRITERIONE ). If M is continuous and exp M < then M E[ ( 2 [ ] )] ( ) is a u.i. martingale. ∞ ∞ E (LÉPINGLE-MÉMIN). If ∆M > 1 and − X A : 1 M c 1 M log 1 M M t = 2 + (( + ∆ s) ( + ∆ s) ∆ s) 〈 〉 s t − ≤ p p has a compensator A such that [exp(A )] < then (M) is a u.i. mar- tingale. E ∞ ∞ E (PROTTER-SHIMBO). If M is locally square integrable such that ∆M > 1 and − exp 1 M c M d < E[ ( 2 + )] 〈 〉∞ 〈 〉∞ ∞ then (M) is a u.i. martingale. E 64 The Bichteler-Dellacherie Theorem and its connexions to arbitrage 3.7.7 Exercise. Why must the supremum in Kazamaki’s criterion be taken over all bounded stopping times and not just deterministic times? Chapter 4 General stochastic integration We know how to integrate elements of L with respect to semimartingales. How- ever, this set of integrands is not rich enough to prove “martingale representation” theorems nor many formulae regarding semimartingale local time. 4.1 Stochastic integrals with respect to predictable processes For this section we will assume that all semimartingales that appear are null at zero. 4.1.1 Definition. Let X be a special semimartingale with canonical decomposition 2 1/2 2 X = M + A. The H -norm of X is X H2 := [M] L2 + A L2 . Define H := ∞ ∞ X : X is a special semimartingalek withk X Hk2 < k k| | k { k k ∞} Remark. This extends the definition given in the previous chapter. A local martin- gale M is in H2 if and only if M is an L2-martingale. 2 4.1.2 Theorem. (H , H2 ) is a Banach space. k · k n n n 2 n 2 PROOF (SKETCH): If X = M + A is Cauchy in H then M is Cauchy in L and so the sequence has a limit M. As an exercise, read the rest∞ of this proof in the textbook. 2 2 4.1.3 Lemma. If H L and H is bounded and X H then H X H . ∈ ∈ · ∈ PROOF: Suppose that X = M+A is the canonical decomposition. By Theorem 3.5.9, H M is a local martingale. To see that H A is predictable, write it as a limit of Riemann· sums. It is clearly of finite variation.· Also notice H X H2 H u X H2 < . k · k ≤ k k k k ∞ 65 66 General stochastic integration It can be seen that 2 1/2 H X H2 = (H [M]) L2 + H A L2 . k · k k · ∞ k k| · |∞k The integrals on the right hand side are well-defined for any bounded process. Notation. If is a σ-algebra then b denotes the collection of bounded - measurable functions.F Let bL denote theF collection of bounded processes in L.F 2 4.1.4 Definition. Let X = M + A H and H, J b . Define ∈ ∈ P 2 1/2 dX (H, J) := ((H J) [M]) L2 + (H J) A L2 . k − · ∞ k k| − · |∞k Remark. dX is not a metric, but it is a semi-metric. 4.1.5 Theorem. If X H2 then ∈ (i) bL is dense in (b , dX ) n P n 2 (ii) If (H )n 1 bL is Cauchy in (b , dX ) then (H X ) is Cauchy in H . n≥ ⊆ n P n ·n (iii) If dX (H , H) 0 and dX (J , H) 0 with H , J bL and H b then n → n →2 ∈ ∈ P limn H X = limn J X in H . →∞ · →∞ · Remark. Theorem 4.1.5(ii) and (iii) can be more succinctly stated as follows. The 2 mapping (bL, dX ) (H , H2 ) that sends H H X is an isometry and can be → k · k 7→ · extended to (b , dX ). P PROOF (OF (I)): Let M := bL and H := H b : for all " > 0 there is J bL such that dX (H, J) < " . { ∈ P ∈ } Then M is a multiplicative class and H is a monotone vector space. Indeed, H n is a vector space that contains constants (exercise). Suppose that (H ) H and H n H pointwise and H is bounded. By the dominated convergence⊆ theorem ↑ n n0 dX (H , H) 0. Let " > 0 and pick n0 such that dX (H , H) < "/2 and pick J bL → n0 ∈ such that dX (H , J) < "/2. Then dX (H, J) < ", so H H. Therefore H is a monotone vector space. By the monotone class theorem b∈ = bσ(M) H. P ⊆ 2 n 4.1.6 Definition. Let X H and H b . Suppose that (H )n 1 bL is such n ∈ ∈ P ≥ ⊆ that dX (H , H) 0 (there is such a sequence by Theorem 4.1.5(i)). Define the stochastic integral→of H with respect to X to be H X : H2-lim H n X , = n ( ) · →∞ · The limit exists by Theorem 4.1.5(ii) and H X is well-defined (i.e. does not depend on the choice of sequence) by Theorem 4.1.5· (iii). 4.1.7 Theorem. If X H2 then X 2 sup X 2 8 X 2 . E[( ∗ ) ] = E[ t 0 t ] H2 ∈ ∞ ≥ | | ≤ k k 4.1. Stochastic integrals with respect to predictable processes 67 ROOF P : Suppose that X = M + A. Then X ∗ M ∗ + A . By Doob’s maximal 2 2 ∞ ≤ ∞ | |∞ inequality, E[(M ∗ ) ] 4 E[M ] = 4 E[[M] ]. Whence ∞ ≤ ∞ ∞ 2 2 2 (X ∗ ) 2(M ∗ ) + 2( A ) ∞2 ≤ ∞ 2 | |∞ 2 E[(X ∗ ) ] 2 E[(M ∗ ) ] + 2 E[( A ) ] ∞ ≤ ∞ | |∞ 2 8 E[[M] ] + 2 E[( A ) ] ≤ 2 ∞ | |∞ 8 X H2 ≤ k k n 2 nk 4.1.8 Corollary. If X X in H then there is a subsequence (X )k 1 such that nk → ≥ (X X )∗ 0 a.s. − ∞ → 4.1.9 Theorem (Properties of the stochastic integral). Let X , Y H2 and H, K b . (i) (H∈+ K) X = H X∈+PK X (ii) H (X +·Y ) = H · X + H · Y · · · T T (iii) If T is a stopping time then (H X ) = (H1[0,T]) X = H (X ). (iv) ∆(H X ) = H∆X · · · · T T (v) If T is a stopping time then H (X −) = (H X ) −. (vi) If X has finite variation then H· X coincides· with the Lebesgue-Stieltjes in- tegral. · (vii) H (K X ) = (HK) X (viii) If X· is· a local martingale· then H X is an L2-martingale. (ix) [H X , K Y ] = (HK) [X , Y ] · · · · 4.1.10 Exercise. For (ii), show that you can take the same approximating se- quence for dX and dY . n n PROOF (SKETCH): For (iv), suppose that (H )n 1 bL are such that dX (H , H) n n≥ ⊆ n → 0 and (H X H X )∗ 0 a.s. Then limn H ∆X = limn ∆(H X ) = ∆(H X ). It follows· that− · ∞ → →∞ →∞ · · n ∆(H X ) lim H 1 X 0 1 X 0 a.s. n ∆ = = · ∆ = 6 X 6 →∞ ∆ n Show that this implies that limn H ∆X = H∆X . For (ix), it is enough to show→∞ that [H X , Y ] = H [X , Y ]. Again suppose n n · n · (H )n 1 bL are such that dX (H , H) 0 and (H X H X )∗ 0 a.s. Since Y is locally≥ ⊆ bounded we can assume that→ it is bounded.· − · ∞ → − n n [H X , Y ] = H [X , Y ] H [X , Y ] n · n· → n · n [H X , Y ] = (H X )Y (H X ) Y Y (H X ) · · − · − · − − · · (H X )Y (H X ) Y Y (H X ) → · − · − · − − · · (This proof may require some further explanation.) 68 General stochastic integration 4.1.11 Definition. A property (P) holds prelocally for X if there exist a sequence T of stopping times with T such that (P) holds for X Tn X 1 ( n)n 1 n − = [0,Tn) + X 1≥ (with the convention that↑ ∞ ). Tn [Tn, ) = − ∞ ∞− ∞ 4.1.12 Theorem. Let X be a semimartingale with X0 = 0. Then X is prelocally an element of H2. PROOF: Note that X is not necessarily as special semimartingale. Write X = M + A where M is a local martingale and A is of finite variation (cf. the Bichteler- Tn Dellacherie Theorem). Let Tn := inf t : Mt > n or A t > n and define Y = X −. { | | | | } Tn Tn Tn Yt∗ = (X −)∗t (M −)∗t + (A −)∗t 2n ≤ ≤ It follows that Y is a special semimartingale. We may assume, by further local- ization, that [Y ] is bounded. (Indeed, take Rn = inf t : [Y ]t > n and note that Y Rn n Y 3n.) Suppose that Y N B where{ N is a local} martingale [ ]∗ +(∆ Rn )∗ = + and B is≤ predictable.≤ We have seen that B is locally bounded (homework), so we may assume that B H2 by localizing| once| more. Again from the homework, 2 E[[Y ] ] = E[[N] ]∈ + E[[B] ], so E[[N] ] < and hence N H as well. ∞ ∞ ∞ ∞ ∞ ∈ 2 4.1.13 Definition. Let X H with canonical decomposition X = M + A. We say 2 that H is (H , X )-integrable∈ if ∈ P Z Z 2 ∞ 2 ∞ E Hs d[M]s + E Hs d A s < . 0 0 | | | | ∞ 2 2 4.1.14 Theorem. If X H and H is (H , X )-integrable then the sequence ∈ ∈ P 2 ((H1 H n) X )n 1 is a Cauchy sequence in H . | |≤ · ≥ n n n PROOF: Let H := H1 H n. Clearly H b , so H X is well-defined. | |≤ ∈ P · n m n m H X H X H2 = dX (H , H ) k · − · k n m 2 1/2 n m 2 = ((H H ) [M]) L2 + (H H ) A L2 k − · k k| − · |∞k 0 by the dominated convergence theorem. → 2 2 4.1.15 Definition. If X H and H is (H , X )-integrable then define H X 2 ∈ ∈ P · to be the limit in H of ((H1 H n) X )n 1, which exists by Theorem 4.1.14. Yet otherwise said, | |≤ · ≥ 2 H X : H -lim H1 H n X . = n ( ) | |≤ · →∞ · For any semimartingale X and H , we say that H X exists if there is a sequence ∈ P · 2 of stopping times (Tn)n 1 that witnesses X is prelocally in H and such that H is 2 Tn ≥ Tn (H , X −)-integrable for all n. In this case we define H X to be H (X −) on · · [0, Tn) and we say that H L(X ), the set of X -integrable processes. ∈ 4.1. Stochastic integrals with respect to predictable processes 69 Remark. Tm Tn Tm (i) H X is well-defined because H (X −) = (H X −) − for n > m (cf. Theo- rem· 4.1.9). By similar reasoning,· H X does not· depend on the prelocalizing sequence chosen. · (ii) If H (b )loc then H L(X ) for any semimartingale X . (iii) Notice∈ thatP if H L(X )∈then H X is a semimartingale. (Prove as an exercise ∈ · that if C is the collection of semimartingales then C = Cloc = Cpreloc.) (iv) Warning: We have not defined H X = H M + H A when X = M + A. If it Is it true? is true we will need to prove it as· a theorem.· · 4.1.16 Theorem. Let X and Y be semimartingales. (i) L(X ) is a vector space and Λ : L(X ) special semimartingales : H H X is a linear map. → { } 7→ · (ii) ∆(H X ) = H∆X for H L(X ). · T ∈ T (iii) (H X ) = (H1[0,T]) X = H (X ) for H L(X ) and stopping times T. R t (iv) If X· is of finite variation· and· H d X ∈< then H X coincides with the 0 s s Lebesgue-Stieltjes integral for all| H| | L(|X ).∞ · (v) Let H, K . If K L(X ) then H ∈L(K X ) if and only if HK L(X ), and in this case∈ P(HK) X∈= H (K X ). ∈ · ∈ (vi) [H X , K Y ] = (HK· ) [X ,· Y ] ·for H L(X ) and K L(Y ). 2 (vii) X · H if· and only if sup· H ∈X 2 < .∈ In this case, H u 1 ( )∗ L ∈ k k ≤ k · ∞k ∞ 1/2 X H2 sup (H X )∗ L2 + 2 [X ] L2 5 X H2 k k ≤ H u 1 k · ∞k k ∞ k ≤ k k k k ≤ and X H2 3 sup (H X )∗ L2 9 X H2 . k k ≤ H u 1 k · ∞k ≤ k k k k ≤ PROOF (OF (VII)): To be added. 4.1.17 Exercise. Suppose that H and X is of finite variation and also that R t H d X < a.s. for all t (this∈ isP the Lebesgue-Stieltjes integral). Is it true 0 s s that| H| | L|(X )?∞ Note that this is not the same statement as Theorem 4.1.16(iv). ∈ 4.1.18 Theorem. Suppose that X is a semimartingale and H L(X ) under P. If then H L X under and H X H X . ∈ Q P ( ) Q P = Q ∈ · · PROOF (SKETCH): Let (Tn)n 1 witness that H L(X ) under P, so that Tn P-a.s. d 2 Tn ≥ ∈ P Q ↑ ∞ and H is H , X -integrable for all n. Let Zt : t , Sn : inf t : Zt > ( −) = E [ d P ] = Rn |F { | | n and Rn := Tn Sn. Since Z − n one can show that } ∧ | | ≤ Rn 1/2 X 2 sup H X 2 2 Q X H (Q) ( )∗ L (Q) + E [[ −] ] k k ≤ H u 1 k · ∞k ∞ k k ≤ sup Z H X 2 1/2 2 P Z X Rn 1/2 = EP[ Rn (( )∗ ) ] + E [ Rn [ −] ] H u 1 · ∞ ∞ k k ≤ 70 General stochastic integration 5 n X 2 p H (P) ≤ k k (The last inequality is an exercise. There was a hint but I missed it.) This implies 2 2 Rn that X H (Q). The proof that H is (H , X −)-integrable under Q is similar. ∈ 4.1.19 Theorem. Let M be a local martingale and H be locally bounded. Then H L(M) and H M is a local martingale. ∈ P ∈ · PROOF: We need the following lemma. 2 4.1.20 Lemma. Suppose that X = M + A, where M H and A is of integrable ∈ n variation. If H then there is a sequence (Hn)n 1 bL such that H M H M 2 n in H and E[ H∈ P H A ] 0. ≥ ∈ · → · | − | · | |∞ → ˜ 2 1/2 PROOF: Let δX (H, J) := E[(H J) [M] ] + E[ H J A ] for all H, J . − · ∞ | − |˜ · | |∞ ∈ P Let H = H : for all " > 0 there is J bL such that δX (H, J) < " . One can see that H{ is∈ a monotone P vector space that∈ contains bL. It follows that }b H. P ⊆ 2 By localization we may assume that H b and M = M 0 + A, where M 0 H and A is of integrable variation (this follows∈ fromP the fundamental theorem of∈ local martingales and a problem from the exam). By the lemma there is (Hn)n 1 bL ˜ n n n n ≥ ∈ such that δX (H , H) 0. We know that H M = H M 0 + H A is a martingale →n 2 · n· · for each n. Finally, H M 0 H M 0 in H and E[ H H A ] 0 imply that · → · | − | · | |∞ → H M 0 and H A are martingales. · · 4.1.21 Theorem (Ansel-Stricker). Let M be a local martingale and H L(M). Then H M is a local martingale if and only if there stopping times Tn ∈ and · Tn ↑ ∞ (θn)n 1 nonpositive integrable random variables such that (H∆M)t θn for all t and≥ all n. ≥ Remark. If H is locally bounded and T is such that M is integrable and H Tn is n T∗n bounded by n then one can take θ : 2nM in the theorem to show that H M n = T∗n is a local martingale. − · ROOF P : If H M is a local martingale then there is are Tn such that (H M)∗T · ↑ ∞ · n is integrable, and one may take θn := 2(H M)∗T to see that the condition is − · n necessary. Proof of sufficiency may be found in the textbook. 4.1.22 Theorem. Let X = M + A be a special semimartingale and H L(X ). If H X is a special semimartingale then H L(M) L(A), H M is a local martingale,∈ and· H A is predictable and of finite variation.∈ In∩ this case· H X = H M + H A is the canonical· decomposition of H X . · · · · 4.1.23 Theorem (Dominated convergence). Let X be a semimartingale, G m m n L(X ), H such that H G, and assume that H H pointwise. Then∈ m n H , H L(∈X ) Pfor all m and| H | ≤X H X u.c.p. → ∈ · → · 4.1. Stochastic integrals with respect to predictable processes 71 2 T ROOF n P : Suppose that Tn are such that G is (H , X −)-integrable for all n. If Tn ↑ ∞ X − = M + A is the canonical decomposition then m 2 1/2 m 2 1/2 2 1/2 2 1/2 E[(H ) [M] ] + E[ H A ] E[G [M] ] + E[ G A ] < · ∞ | | · | |∞ ≤ · ∞ | | · | |∞ ∞ m 2 Tn Therefore H are all (H , X −)-integrable and hence so is H by the dominated convergence theorem. For t0 fixed m Tn 2 m Tn 2 sup (H H) X − L2 8 (H H) X − H2 0 k t t0 | − · |k ≤ k − · k → ≤ m Tn Tn by the dominated convergence theorem. Whence H X − H X − uniformly · → · on [0, t0] in probability. Given " > 0 let n and m be such that P[Tn < t0] < " and m Tn m P[((H H) X −)∗t > δ] < ". Whence P[((H H) X )∗t > δ] < 2". − · 0 − · 0 t 4.1.24 Theorem. Let X be a semimartingale. There exists Q P such that X H2 for all t and d Q is bounded. ∼ ∈ (Q) d P 4.1.25 Example (Emery). Suppose that T Exponential(1), P[U = 1] = P[U = 1 1 , and U and T are independent. Let∼X : U1 and let the filtration be ] = 2 = t T −the natural filtration of X , so that X is an H2 martingale≥ of finite variation and X 1 . Let H : 1 1 , which is predictable because it is left continuous. [ ]t = t T t = t t>0 ≥ 2 But H is not (H , X )-integrable. 2 1 E[H [X ] ] = E 2 = · ∞ T ∞ Is H L(X )? We do know that H X exists as a pathwise Lebesgue-Stieltjes integral,∈ namely H X U 1 . We· have H X for all t (because the ( )t = T t T E[ t ] = problem is around zero· and not≥ at ). It follows| · in| particular∞ that H X is not a martingale. ∞ · Claim. If S is a finite stopping time with P[S > 0] > 0 then E[ (H X )S ] = . | · | ∞ Indeed, we have H X 1 1 . The hard part is to show that there is " > 0 )S = T S T | · | ≥ such that “ T " implies S T ”. If this is the case then E[ (H X )S ] = 1 1 { ≤. } { ≥ } | · | E[ T T "] = ≤ ∞ (Perhaps use T " = T ", S < T T ", S T and 1S 2 (H , X )-integrable, 68 indistinguishable,5 H2-norm, 65 infinitely divisible distribution, 18 S, 24 integrable processes, 68 integrable variation, 54 adapted,5 intensity, 10 additive process, 12 intrinsic Lévy process, 12 announcing sequence, 54 arbitrage, 44 Jacod’s countable expansion theorem, arrival rate, 10 25 Brownian motion, 11 Lévy’s theorem, 42 Lévy-Khintchine exponent, 18 càdlàg,5 Lévy measure, 16 càglàd,5 Lévy process, 12 canonical decomposition, 58 little investment, 44 class D, 52 local martingale, 20 classical semimartingale, 43 localized class, 20 closable,8 localizing sequence, 20 compensator, 57 continuous part of quadratic varia- modification,5 tion, 32 monotone vector space,6 counting process,9 NA, 51 début time,6 natural process, 54 decomposable process, 26 NFLVR, 44 NFLVR+LI, 44 events before time T,5 no arbitrage, 51 events strictly before time T,5 no free lunch with vanishing risk, 44 explosion time,9 no unbounded profit with bounded risk, 51 Feller process, 13 NUPBR, 51 Fisk-Stratonovich integral, 39 FLVR, 44 optional σ-algebra,6 FLVR+LI, 44 optional process,6 free lunch with vanishing risk, 44 orthogonal, 59 hitting time,6 Poisson process,9 72 Index 73 polarization identity, 31 predictable, 54 predictable variation, 57 prelocally, 68 price process, 44 process of jumps,5 progressive σ-algebra,7 progressive process,7 progressively measurable,7 purely discontinuous, 59 quadratic pure jump semimartingale, 32 quadratic variation, 31 quasimartingale, 56 random time,4 semimartingale, 24 semimartingale characteristics, 60 simple predictable process, 24 spatially homogeneous, 13 special semimartingale, 58 square integrable martingale,9 stable process, 12 stable under stopping, 20 standard Brownian motion, 11 stochastic integral, 24, 28, 66 stochastic integral process, 28 stochastic process,5 stopped process,9 stopping time,4 strategy, 44 subordinator, 12 total semimartingale, 24 total variation process, 21 totally inaccessible, 54 u.c.p., 27 u.i.,7 uniformly integrable,7 uniformly on compacts in probabil- ity, 27 usual conditions,4 variation, 56 0. Taylor’s theorem states that, for x, y [ K, K], ∈ − 1 2 f (y) f (x) = f 0(x)(y x) + f 00(x)(y x) + R(x, y), − − 2 − 2.7. Itô’s formula 37 t. One can rewrite the above in terms of Itô’s formula for X . (More perspicaciously, for fixed t, for each ω there is k large k enough so that Y (ω, t) = X (ω, t) and the differentials are equal too.) Stratonovich integral of is defined for semimartingales R t R t 1 X and Y by Y X : Y dX : Y dX X , Y c. ( )t = 0 s s = 0 s s + 2 [ ]t − ◦ − ◦ − 2.7.4 Theorem. 3 (i) If X is a semimartingale and f C (R) ∈ Z t X f (X t ) f (X0) = f 0(Xs ) dXs + (f (Xs) f (Xs ) f 0(Xs )∆Xs). − 0+ − ◦ 0