STOCHASTIC PROCESSES AND APPLICATIONS
G.A. Pavliotis Department of Mathematics Imperial College London
November 11, 2015 2 Contents
1 Stochastic Processes 1 1.1 Definition of a Stochastic Process ...... 1 1.2 StationaryProcesses ...... 3 1.3 BrownianMotion ...... 10 1.4 Examples of Stochastic Processes ...... 13 1.5 The Karhunen-Lo´eve Expansion ...... 16 1.6 Discussion and Bibliography ...... 20 1.7 Exercises ...... 21
2 Diffusion Processes 27 2.1 Examples of Markov processes ...... 27 2.2 The Chapman-Kolmogorov Equation ...... 30 2.3 The Generator of a Markov Processes ...... 34 2.4 Ergodic Markov processes ...... 37 2.5 The Kolmogorov Equations ...... 39 2.6 Discussion and Bibliography ...... 46 2.7 Exercises ...... 47
3 Introduction to SDEs 49 3.1 Introduction...... 49 3.2 The Itˆoand Stratonovich Stochastic Integrals ...... 52 3.3 SolutionsofSDEs...... 57 3.4 Itˆo’sformula...... 59 3.5 ExamplesofSDEs ...... 65 3.6 Lamperti’s Transformation and Girsanov’s Theorem ...... 68 3.7 Linear Stochastic Differential Equations ...... 70 3.8 Discussion and Bibliography ...... 73 3.9 Exercises ...... 74
4 The Fokker-Planck Equation 77 4.1 Basic properties of the Fokker-Planck equation ...... 77 4.2 Examples of the Fokker-Planck Equation ...... 81
i 4.3 Diffusion Processes in One Dimension ...... 87 4.4 The Ornstein-Uhlenbeck Process ...... 90 4.5 TheSmoluchowskiEquation ...... 96 4.6 Reversible Diffusions ...... 101 4.7 Eigenfunction Expansions ...... 106 4.8 MarkovChainMonteCarlo...... 108 4.9 Reduction to a Schr¨odinger Operator ...... 110 4.10 Discussion and Bibliography ...... 114 4.11Exercises ...... 118
Appendix A Frequently Used Notation 123
Appendix B Elements of Probability Theory 125 B.1 Basic Definitions from Probability Theory ...... 125 B.2 RandomVariables...... 127 B.3 Conditional Expecation ...... 131 B.4 The Characteristic Function ...... 132 B.5 Gaussian Random Variables ...... 133 B.5.1 Gaussian Measures in Hilbert Spaces ...... 135 B.6 Types of Convergence and Limit Theorems ...... 138 B.7 Discussion and Bibliography ...... 139
Index 141
Bibliography 145
ii Chapter 1
Introduction to Stochastic Processes
In this chapter we present some basic results from the theory of stochastic processes and investigate the properties of some of the standard continuous-time stochastic processes. In Section 1.1 we give the definition of a stochastic process. In Section 1.2 we present some properties of stationary stochastic processes. In Section 1.3 we introduce Brownian motion and study some of its properties. Various examples of stochastic processes in continuous time are presented in Section 1.4. The Karhunen-Loeve expansion, one of the most useful tools for representing stochastic processes and random fields, is presented in Section 1.5. Further discussion and bibliographical comments are presented in Section 1.6. Section 1.7 contains exercises.
1.1 Definition of a Stochastic Process
Stochastic processes describe dynamical systems whose time-evolution is of probabilistic nature. The pre- cise definition is given below.1
Definition 1.1 (stochastic process). Let T be an ordered set, (Ω, , P) a probability space and (E, ) a F G measurable space. A stochastic process is a collection of random variables X = X ; t T where, for { t ∈ } each fixed t T , X is a random variable from (Ω, , P) to (E, ). Ω is known as the sample space, where ∈ t F G E is the state space of the stochastic process Xt.
The set T can be either discrete, for example the set of positive integers Z+, or continuous, T = R+. The state space E will usually be Rd equipped with the σ-algebra of Borel sets. A stochastic process X may be viewed as a function of both t T and ω Ω. We will sometimes write ∈ ∈ X(t), X(t,ω) or X (ω) instead of X . For a fixed sample point ω Ω, the function X (ω) : T E is t t ∈ t → called a (realization, trajectory) of the process X. Definition 1.2 (finite dimensional distributions). The finite dimensional distributions (fdd) of a stochastic k process are the distributions of the E -valued random variables (X(t1), X(t2),...,X(tk)) for arbitrary positive integer k and arbitrary times t T, i 1,...,k : i ∈ ∈ { }
F (x)= P(X(ti) 6 xi, i = 1,...,k) with x = (x1,...,xk). 1The notation and basic definitions from probability theory that we will use can be found in Appendix B.
1 From experiments or numerical simulations we can only obtain information about the finite dimensional distributions of a process. A natural question arises: are the finite dimensional distributions of a stochastic process sufficient to determine a stochastic process uniquely? This is true for processes with continuous paths 2, which is the class of stochastic processes that we will study in these notes.
Definition 1.3. We say that two processes Xt and Yt are equivalent if they have same finite dimensional distributions.
Gaussian stochastic processes
A very important class of continuous-time processes is that of Gaussian processes which arise in many applications.
Definition 1.4. A one dimensional continuous time Gaussian process is a stochastic process for which E = R and all the finite dimensional distributions are Gaussian, i.e. every finite dimensional vector (X , X ,...,X ) is a ( , K ) random variable for some vector and a symmetric nonnegative t1 t2 tk N k k k definite matrix Kk for all k = 1, 2,... and for all t1,t2,...,tk.
From the above definition we conclude that the finite dimensional distributions of a Gaussian continuous- time stochastic process are Gaussian with probability distribution function
n/2 1/2 1 1 γ (x) = (2π)− (detK )− exp K− (x ), x , k,Kk k −2 k − k − k where x = (x1,x2,...xk). It is straightforward to extend the above definition to arbitrary dimensions. A Gaussian process x(t) is characterized by its mean m(t) := Ex(t) and the covariance (or autocorrelation) matrix
C(t,s)= E x(t) m(t) x(s) m(s) . − ⊗ − Thus, the first two moments of a Gaussian process are sufficien t for a complete characterization of the process. It is not difficult to simulate Gaussian stochastic processes on a computer. Given a random number generator that generates (0, 1) (pseudo)random numbers, we can sample from a Gaussian stochastic pro- N cess by calculating the square root of the covariance. A simple algorithm for constructing a skeleton of a continuous time Gaussian process is the following:
Fix ∆t and define t = (j 1)∆t, j = 1,...N. • j − N N N N N N Set Xj := X(tj) and define the Gaussian random vector X = Xj . Then X ( , Γ ) • j=1 ∼N N N with = ( (t1),... (tN )) and Γij = C(ti,tj).
2In fact, all we need is the stochastic process to be separable See the discussion in Section 1.6.
2 Then XN = N + Λ (0,I) with ΓN = ΛΛT . • N We can calculate the square root of the covariance matrix C either using the Cholesky factorization, via the spectral decomposition of C, or by using the singular value decomposition (SVD).
Examples of Gaussian stochastic processes
Random Fourier series: let ξ , ζ (0, 1), i = 1,...N and define • i i ∼N N X(t)= (ξj cos(2πjt)+ ζj sin(2πjt)) . j=1
Brownian motion is a Gaussian process with m(t) = 0, C(t,s) = min(t,s). • Brownian bridge is a Gaussian process with m(t) = 0, C(t,s) = min(t,s) ts. • − The Ornstein-Uhlenbeck process is a Gaussian process with m(t) = 0, C(t,s) = λ e α t s with • − | − | α, λ> 0.
1.2 Stationary Processes
In many stochastic processes that appear in applications their statistics remain invariant under time transla- tions. Such stochastic processes are called stationary. It is possible to develop a quite general theory for stochastic processes that enjoy this symmetry property. It is useful to distinguish between stochastic pro- cesses for which all finite dimensional distributions are translation invariant (strictly stationary processes) and processes for which this translation invariance holds only for the first two moments (weakly stationary processes).
Strictly Stationary Processes
Definition 1.5. A stochastic process is called (strictly) stationary if all finite dimensional distributions are invariant under time translation: for any integer k and times t T , the distribution of (X(t ), X(t ),...,X(t )) i ∈ 1 2 k is equal to that of (X(s+t ), X(s+t ),...,X(s+t )) for any s such that s+t T for all i 1,...,k . 1 2 k i ∈ ∈ { } In other words,
P(X A , X A ...X A )= P(X A , X A ...X A ), s T. t1+s ∈ 1 t2+s ∈ 2 tk+s ∈ k t1 ∈ 1 t2 ∈ 2 tk ∈ k ∀ ∈
Example 1.6. Let Y0, Y1,... be a sequence of independent, identically distributed random variables and consider the stochastic process Xn = Yn. Then Xn is a strictly stationary process (see Exercise 1). Assume furthermore that EY = < + . Then, by the strong law of large numbers, Equation (B.26), we have that 0 ∞ N 1 N 1 1 − 1 − X = Y EY = , N j N j → 0 j=0 j=0 3 almost surely. In fact, the Birkhoff ergodic theorem states that, for any function f such that Ef(Y ) < + , 0 ∞ we have that N 1 1 − lim f(Xj)= Ef(Y0), (1.1) N + N → ∞ j=0 almost surely. The sequence of iid random variables is an example of an ergodic strictly stationary processes.
We will say that a stationary stochastic process that satisfies (1.1) is ergodic. For such processes we can calculate expectation values of observable, Ef(Xt) using a single sample path, provided that it is long enough (N 1). ≫
Example 1.7. Let Z be a random variable and define the stochastic process Xn = Z, n = 0, 1, 2,... . Then Xn is a strictly stationary process (see Exercise 2). We can calculate the long time average of this stochastic process: N 1 N 1 1 − 1 − X = Z = Z, N j N j=0 j=0 which is independent of N and does not converge to the mean of the stochastic processes EXn = EZ (assuming that it is finite), or any other deterministic number. This is an example of a non-ergodic processes.
Second Order Stationary Processes
Let Ω, , P be a probability space. Let X , t T (with T = R or Z) be a real-valued random process F t ∈ on this probability space with finite second moment, E X 2 < + (i.e. X L2(Ω, P) for all t T ). | t| ∞ t ∈ ∈ Assume that it is strictly stationary. Then,
E(X )= EX , s T, (1.2) t+s t ∈ from which we conclude that EXt is constant and
E((X )(X )) = E((X )(X )), s T, (1.3) t1+s − t2+s − t1 − t2 − ∈ implies that the covariance function depends on the difference between the two times, t and s:
C(t,s)= C(t s). − This motivates the following definition.
Definition 1.8. A stochastic process X L2 is called second-order stationary, wide-sense stationary or t ∈ weakly stationary if the first moment EX is a constant and the covariance function E(X )(X ) t t − s − depends only on the difference t s: − EX = , E((X )(X )) = C(t s). t t − s − −
The constant is the expectation of the process Xt. Without loss of generality, we can set = 0, since if EX = then the process Y = X is mean zero. A mean zero process is called a centered process. t t t − The function C(t) is the covariance (sometimes also called autocovariance) or the autocorrelation function
4 E E 2 of the Xt. Notice that C(t) = (XtX0), whereas C(0) = Xt , which is finite, by assumption. Since we have assumed that X is a real valued process, we have that C(t)= C( t), t R. t − ∈ Let now Xt be a strictly stationary stochastic process with finite second moment. The definition of strict stationarity implies that EX = , a constant, and E((X )(X )) = C(t s). Hence, a strictly t t − s − − stationary process with finite second moment is also stationary in the wide sense. The converse is not true, in general. It is true, however, for Gaussian processes: since the first two moments of a Gaussian process are sufficient for a complete characterization of the process, a Gaussian stochastic process is strictly stationary if and only if it is weakly stationary.
Example 1.9. Let Y0, Y1,... be a sequence of independent, identically distributed random variables and consider the stochastic process Xn = Yn. From Example 1.6 we know that this is a strictly stationary process, irrespective of whether Y is such that EY 2 < + . Assume now that EY = 0 and EY 2 = σ2 < 0 0 ∞ 0 0 + . Then X is a second order stationary process with mean zero and correlation function R(k)= σ2δ . ∞ n k0 Notice that in this case we have no correlation between the values of the stochastic process at different times n and k.
Example 1.10. Let Z be a single random variable and consider the stochastic process Xn = Z, n = 0, 1, 2,... . From Example 1.7 we know that this is a strictly stationary process irrespective of whether E Z 2 < + or not. Assume now that EZ = 0,EZ2 = σ2. Then X becomes a second order stationary | | ∞ n process with R(k) = σ2. Notice that in this case the values of our stochastic process at different times are strongly correlated.
We will see later in this chapter that for second order stationary processes, ergodicity is related to fast decay of correlations. In the first of the examples above, there was no correlation between our stochastic processes at different times and the stochastic process is ergodic. On the contrary, in our second example there is very strong correlation between the stochastic process at different times and this process is not ergodic. Continuity properties of the covariance function are equivalent to continuity properties of the paths of 2 Xt in the L sense, i.e. 2 lim E Xt+h Xt = 0. h 0 | − | → Lemma 1.11. Assume that the covariance function C(t) of a second order stationary process is continuous at t = 0. Then it is continuous for all t R. Furthermore, the continuity of C(t) is equivalent to the 2 ∈ continuity of the process Xt in the L -sense.
Proof. Fix t R and (without loss of generality) set EX = 0. We calculate: ∈ t C(t + h) C(t) 2 = E(X X ) E(X X ) 2 = E ((X X )X ) 2 | − | | t+h 0 − t 0 | | t+h − t 0 | 6 E(X )2E(X X )2 0 t+h − t = C(0) EX2 + EX2 2E(X X ) t+h t − t t+h = 2C(0)(C(0) C(h)) 0, − → as h 0. Thus, continuity of C( ) at 0 implies continuity for all t. → 5 Assume now that C(t) is continuous. From the above calculation we have
E X X 2 = 2(C(0) C(h)), (1.4) | t+h − t| − which converges to 0 as h 0. Conversely, assume that X is L2-continuous. Then, from the above → t equation we get limh 0 C(h)= C(0). → Notice that form (1.4) we immediately conclude that C(0) >C(h), h R. ∈ The Fourier transform of the covariance function of a second order stationary process always exists. This enables us to study second order stationary processes using tools from Fourier analysis. To make the link between second order stationary processes and Fourier analysis we will use Bochner’s theorem, which applies to all nonnegative functions.
Definition 1.12. A function f(x) : R R is called nonnegative definite if → n f(t t )c c¯ > 0 (1.5) i − j i j i,j =1 for all n N, t ,...t R, c ,...c C. ∈ 1 n ∈ 1 n ∈ Lemma 1.13. The covariance function of second order stationary process is a nonnegative definite function.
c n Proof. We will use the notation Xt := i=1 Xti ci. We have. n n C(t t )c c¯ = EX X c c¯ i − j i j ti tj i j i,j =1 i,j =1 n n = E X c X c¯ = E XcX¯ c ti i tj j t t i=1 j=1 = E Xc 2 > 0. | t |
Theorem 1.14. [Bochner] Let C(t) be a continuous positive definite function. Then there exists a unique nonnegative measure ρ on R such that ρ(R)= C(0) and
C(t)= eiωt ρ(dω) t R. (1.6) R ∀ ∈ Let Xt be a second order stationary process with autocorrelation function C(t) whose Fourier transform is the measure ρ(dω). The measure ρ(dω) is called the spectral measure of the process Xt. In the following we will assume that the spectral measure is absolutely continuous with respect to the Lebesgue measure on R with density S(ω), i.e. ρ(dω)= S(ω)dω. The Fourier transform S(ω) of the covariance function is called the spectral density of the process:
1 ∞ itω S(ω)= e− C(t) dt. (1.7) 2π −∞ 6 From (1.6) it follows that that the autocorrelation function of a mean zero, second order stationary process is given by the inverse Fourier transform of the spectral density:
C(t)= ∞ eitωS(ω) dω. (1.8) −∞
The autocorrelation function of a second order stationary process enables us to associate a timescale to Xt, the correlation time τcor: 1 1 τ = ∞ C(τ) dτ = ∞ E(X X ) dτ. cor C(0) E(X2) τ 0 0 0 0 The slower the decay of the correlation function, the larger the correlation time is. Notice that when the correlations do not decay sufficiently fast so that C(t) is not integrable, then the correlation time will be infinite.
Example 1.15. Consider a mean zero, second order stationary process with correlation function
α t C(t)= C(0)e− | | (1.9)
D where α> 0. We will write C(0) = α where D> 0. The spectral density of this process is:
+ 1 D ∞ iωt α t S(ω) = e− e− | | dt 2π α −∞ 0 + 1 D iωt αt ∞ iωt αt = e− e dt + e− e− dt 2π α 0 −∞ 1 D 1 1 = + 2π α iω + α iω + α − D 1 = . π ω2 + α2 This function is called the Cauchy or the Lorentz distribution. The correlation time is (we have that R(0) = D/α) ∞ αt 1 τcor = e− dt = α− . 0 A real-valued Gaussian stationary process defined on R with correlation function given by (1.9) is called the stationary Ornstein-Uhlenbeck process. We will study this stochastic process in detail in later chapters.
The Ornstein-Uhlenbeck process Xt can be used as a model for the velocity of a Brownian particle. It is of interest to calculate the statistics of the position of the Brownian particle, i.e. of the integral (we assume that the Brownian particle starts at 0) t Zt = Ys ds, (1.10) 0 The particle position Zt is a mean zero Gaussian process. Set α = D = 1. The covariance function of Zt is
min(t,s) max(t,s) t s E(Z Z ) = 2min(t,s)+ e− + e− e−| − | 1. (1.11) t s − −
7 Ergodic properties of second-order stationary processes
Second order stationary processes have nice ergodic properties, provided that the correlation between values of the process at different times decays sufficiently fast. In this case, it is possible to show that we can calculate expectations by calculating time averages. An example of such a result is the following.
Proposition 1.16. Let X > be a second order stationary process on a probability space (Ω, , P) with { t}t 0 F mean and covariance C(t), and assume that C(t) L1(0, + ). Then ∈ ∞ 2 1 T lim E Xs ds = 0. (1.12) T + T − → ∞ 0 For the proof of this result we will first need the following result, which is a property of symmetric functions. Lemma 1.17. Let R(t) be an integrable symmetric function. Then
T T T C(t s) dtds = 2 (T s)C(s) ds. (1.13) − − 0 0 0 Proof. We make the change of variables u = t s, v = t + s. The domain of integration in the t, s − variables is [0, T ] [0, T ]. In the u, v variables it becomes [ T, T ] [ u , 2T u ]. The Jacobian of the × − × | | − | | transformation is ∂(t,s) 1 J = = . ∂(u, v) 2 The integral becomes
T T T 2T u R(t s) dtds = −| | R(u)J dvdu 0 0 − T u − | | T = (T u )R(u) du T − | | − T = 2 (T u)R(u) du, − 0 where the symmetry of the function C(u) was used in the last step.
Proof of Theorem 1.16. We use Lemma (1.17) to calculate: 2 2 1 T 1 T E X ds = E (X ) ds T s − T 2 s − 0 0 T T 1 = E (Xt )( Xs ) dtds T 2 − − 0 0 1 T T = C(t s) dtds T 2 − 0 0 2 T = (T u)C(u) du T 2 − 0 2 + u 2 + 6 ∞ 1 C(u) du 6 ∞ C(u) du 0, T 0 − T T 0 → 8 using the dominated convergence theorem and the assumption C( ) L1(0, + ). Assume that = 0 ∈ ∞ and define + D = ∞ C(t) dt, (1.14) 0 which, from our assumption on C(t), is a finite quantity. 3 The above calculation suggests that, for t 1, ≫ we have that t 2 E X(t) dt 2Dt. ≈ 0 This implies that, at sufficiently long times, the mean square displacement of the integral of the ergodic second order stationary process Xt scales linearly in time, with proportionality coefficient 2D. Let now Xt be the velocity of a (Brownian) particle. The particle position Zt is given by (1.10). From our calculation above we conclude that E 2 Zt = 2Dt. where ∞ ∞ D = C(t) dt = E(XtX0) dt (1.15) 0 0 is the diffusion coefficient. Thus, one expects that at sufficiently long times and under appropriate assump- tions on the correlation function, the time integral of a stationary process will approximate a Brownian motion with diffusion coefficient D. The diffusion coefficient is an example of a transport coefficient and (1.15) is an example of the Green-Kubo formula: a transport coefficient can be calculated in terms of the time integral of an appropriate autocorrelation function. In the case of the diffusion coefficient we need to calculate the integral of the velocity autocorrelation function. We will explore this topic in more detail in Chapter ??.
Example 1.18. Consider the stochastic processes with an exponential correlation function from Exam- ple 1.15, and assume that this stochastic process describes the velocity of a Brownian particle. Since C(t) L1(0, + ) Proposition 1.16 applies. Furthermore, the diffusion coefficient of the Brownian particle ∈ ∞ is given by + ∞ 1 D C(t) dt = C(0)τ − = . c α2 0 Remark 1.19. Let X be a strictly stationary process and let f be such that E(f(X ))2 < + . A calcula- t 0 ∞ tion similar to the one that we did in the proof of Proposition 1.16 enables to conclude that
1 T lim f(Xs) ds = Ef(X0), (1.16) T + T → ∞ 0 2 in L (Ω). In this case the autocorrelation function of Xt is replaced by
C (t)= E f(X ) Ef(X ) f(X ) Ef(X ) . f t − 0 0 − 0 3Notice however that we do not know whether it is nonzero. This requires a separate argument.
9 Setting f = f E f we have: − π T + 1 ∞ lim T Varπ f(Xt) dt = 2 Eπ(f(Xt)f(X0)) dt (1.17) T + T → ∞ 0 0 We calculate
2 1 T 1 T Var f(X ) dt = E f(X ) dt π T t π T t 0 0 T T 1 E = 2 π f(Xt)f(Xs) dtds T 0 0 1 T T =: R (t,s) dtds T 2 f 0 0 2 T = (T s)R (s) ds T 2 − f 0 2 T s = 1 Eπ f(Xs)f(X0) ds, T 0 − T from which (1.16) follows.
1.3 Brownian Motion
The most important continuous-time stochastic process is Brownian motion. Brownian motion is a process with almost surely continuous paths and independent Gaussian increments. A process Xt has independent increments if for every sequence t0 X X , X X ,...,X X t1 − t0 t2 − t1 tn − tn−1 are independent. If, furthermore, for any t , t , s T and Borel set B R 1 2 ∈ ⊂ P(X X B)= P(X X B), t2+s − t1+s ∈ t2 − t1 ∈ then the process Xt has stationary independent increments. Definition 1.20. A one dimensional standard Brownian motion W (t) : R+ R is a real valued stochastic → process with a.s. continuous paths such that W (0) = 0, it has independent increments and for every t>s > 0, the increment W (t) W (s) has a Gaussian distribution with mean 0 and variance t s, i.e. − − the density of the random variable W (t) W (s) is − 1 2 2 x g(x; t,s)= 2π(t s) − exp ; (1.18) − −2(t s) − A standard d-dimensional standard Brownian motion W (t) : R+ Rd is a vector of d independent one- → dimensional Brownian motions: W (t) = (W1(t),...,Wd(t)), 10 2 mean of 1000 paths 5 individual paths 1.5 1 0.5 U(t) 0 −0.5 −1 −1.5 0 0.2 0.4 0.6 0.8 1 t Figure 1.1: Brownian sample paths where Wi(t), i = 1,...,d are independent one dimensional Brownian motions. The density of the Gaussian random vector W (t) W (s) is thus − d/2 x 2 g(x; t,s)= 2π(t s) − exp . − −2(t s) − Brownian motion is also referred to as the Wiener process. If Figure 1.1 we plot a few sample paths of Brownian motion. As we have already mentioned, Brownian motion has almost surely continuous paths. More precisely, it has a continuous modification: consider two stochastic processes X and Y , t T , that are defined on the t t ∈ same probability space (Ω, , P). The process Y is said to be a modification of X if P(X = Y ) = 1 for F t t t t all t T . The fact that there is a continuous modification of Brownian motion follows from the following ∈ result which is due to Kolmogorov. Theorem 1.21. (Kolmogorov) Let X , t [0, ) be a stochastic process on a probability space (Ω, , P). t ∈ ∞ F Suppose that there are positive constants α and β, and for each T > 0 there is a constant C(T ) such that E X X α 6 C(T ) t s 1+β, 0 6 s,t 6 T. (1.19) | t − s| | − | Then there exists a continuous modification Yt of the process Xt. We can check that (1.19) holds for Brownian motion with α = 4 and β = 1 using (1.18). It is possible to prove rigorously the existence of the Wiener process (Brownian motion): Theorem 1.22. (Wiener) There exists an almost surely continuous process Wt with independent increments such and W = 0, such that for each t > 0 the random variable W is (0,t). Furthermore, W is almost 0 t N t surely locally Holder¨ continuous with exponent α for any α (0, 1 ). ∈ 2 11 50−step random walk 1000−step random walk 8 20 6 10 4 0 2 −10 0 −20 −2 −30 −4 −40 −6 −50 0 5 10 15 20 25 30 35 40 45 50 0 100 200 300 400 500 600 700 800 900 1000 a. n = 50 b. n = 1000 Figure 1.2: Sample paths of the random walk of length n = 50 and n = 1000. Notice that Brownian paths are not differentiable. We can construct Brownian motion through the limit of an appropriately rescaled random walk: let X , X ,... be iid random variables on a probability space (Ω, , P) with mean 0 and variance 1. Define 1 2 F the discrete time stochastic process Sn with S0 = 0, Sn = j=1 Xj, n > 1. Define now a continuous time stochastic process with continuous paths as the linearly interpolated, appropriately rescaled random walk: 1 1 W n = S + (nt [nt]) X , t √n [nt] − √n [nt]+1 where [ ] denotes the integer part of a number. Then W n converges weakly, as n + to a one dimen- t → ∞ sional standard Brownian motion. See Figure 1.2. An alternative definition of the one dimensional standard Brownian motion is that of a Gaussian stochas- tic process on a probability space Ω, , P with continuous paths for almost all ω Ω, and finite dimen- F ∈ sional distributions with zero mean and covariance E(W W ) = min(t ,t ). One can then show that ti tj i j Definition 1.20 follows from the above definition. For the d-dimensional Brownian motion we have (see (B.7) and (B.8)) EW (t) = 0 t > 0 ∀ and E (W (t) W (s)) (W (t) W (s)) = (t s)I, (1.20) − ⊗ − − where I denotes the identity matrix. Moreover, E W (t) W (s) = min(t,s)I. (1.21) ⊗ Although Brownian motion has stationary increments, it is not a stationary process itself Brownian motion 12 itself. The probability density of the one dimensional Brownian motion is 1 x2/2t g(x,t)= e− . √2πt We can easily calculate all moments: + n 1 ∞ n x2/2t E(W (t) ) = x e− dx √2πt −∞ 1.3 . . . (n 1)tn/2, n even, = − 0, n odd. In particular, the mean square displacement of Brownian motion grows linearly in time. Brownian motion is invariant under various transformations in time. Proposition 1.23. Let Wt denote a standard Brownian motion in R. Then, Wt has the following properties: 1 > > i. (Rescaling). For each c> 0 define Xt = √c W (ct). Then (Xt, t 0) = (Wt, t 0) in law. ii. (Shifting). For each c> 0 W W ,t > 0 is a Brownian motion which is independent of W , u c+t − c u ∈ [0, c]. iii. (Time reversal). Define Xt = W1 t W1, t [0, 1]. Then (Xt, t [0, 1]) = (Wt, t [0, 1]) in law. − − ∈ ∈ ∈ iv. (Inversion). Let Xt, t > 0 defined by X0 = 0, Xt = tW (1/t). Then (Xt, t > 0) = (Wt, t > 0) in law. The equivalence in the above result holds in law and not in a pathwise sense. The proof of this proposi- tion is left as an exercise. We can also add a drift and change the diffusion coefficient of the Brownian motion: we will define a Brownian motion with drift and variance σ2 as the process Xt = t + σWt. The mean and variance of Xt are EX = t, E(X EX )2 = σ2t. t t − t Notice that Xt satisfies the equation dXt = dt + σ dWt. This is an example of a stochastic differential equation. We will study stochastic differential equations in Chapters 3 and ??. 1.4 Examples of Stochastic Processes We present now a few examples of stochastic processes that appear frequently in applications. 13 The Ornstein-Uhlenbeck process The stationary Ornstein-Uhlenbeck process that was introduced earlier in this chapter can be defined through the Brownian motion via a time change. Lemma 1.24. Let W (t) be a standard Brownian motion and consider the process t 2t V (t)= e− W (e ). Then V (t) is a Gaussian stationary process with mean 0 and correlation function t R(t)= e−| |. (1.22) For the proof of this result we first need to show that time changed Gaussian processes are also Gaussian. Lemma 1.25. Let X(t) be a Gaussian stochastic process and let Y (t) = X(f(t)) where f(t) is a strictly increasing function. Then Y (t) is also a Gaussian process. Proof. We need to show that, for all positive integers N and all sequences of times t , t ,...t the { 1 2 N } random vector Y (t ), Y (t ),...Y (t ) (1.23) { 1 2 N } is a multivariate Gaussian random variable. Since f(t) is strictly increasing, it is invertible and hence, there 1 exist si, i = 1,...N such that si = f − (ti). Thus, the random vector (1.23) can be rewritten as X(s ), X(s ),...X(s ) , { 1 2 N } which is Gaussian for all N and all choices of times s1, s2,...sN . Hence Y (t) is also Gaussian. Proof of Lemma 1.24. The fact that V (t) is a mean zero process follows immediately from the fact that W (t) is mean zero. To show that the correlation function of V (t) is given by (1.22), we calculate t s 2t 2s t s 2t 2s E(V (t)V (s)) = e− − E(W (e )W (e )) = e− − min(e , e ) t s = e−| − |. The Gaussianity of the process V (t) follows from Lemma 1.25 (notice that the transformation that gives 1/2 1 V (t) in terms of W (t) is invertible and we can write W (s)= s V ( 2 ln(s))). Brownian Bridge We can modify Brownian motion so that the resulting processes is fixed at both ends. Let W (t) be a standard one dimensional Brownian motion. We define the Brownian bridge (from 0 to 0) to be the process B = W tW , t [0, 1]. (1.24) t t − 1 ∈ Notice that B0 = B1 = 0. Equivalently, we can define the Brownian bridge to be the continuous Gaussian process B : 0 6 t 6 1 such that { t } EB = 0, E(B B ) = min(s,t) st, s,t [0, 1]. (1.25) t t s − ∈ 14 1.5 1 0.5 0 t B −0.5 −1 −1.5 −2 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 t Figure 1.3: Sample paths and first (blue curve) and second (black curve) moment of the Brownian bridge. Another, equivalent definition of the Brownian bridge is through an appropriate time change of the Brownian motion: t B = (1 t)W , t [0, 1). (1.26) t − 1 t ∈ − Conversely, we can write the Brownian motion as a time change of the Brownian bridge: t W = (t + 1)B , t > 0. t 1+ t We can use the algorithm for simulating Gaussian processes to generate paths of the Brownian bridge process and to calculate moments. In Figure 1.3 we plot a few sample paths and the first and second moments of Brownian bridge. Fractional Brownian Motion The fractional Brownian motion is a one-parameter family of Gaussian processes whose increments are correlated. Definition 1.26. A (normalized) fractional Brownian motion W H , t > 0 with Hurst parameter H (0, 1) t ∈ is a centered Gaussian process with continuous sample paths whose covariance is given by 1 E(W H W H)= s2H + t2H t s 2H . (1.27) t s 2 − | − | The Hurst exponent controls the correlations between the increments of fractional Brownian motion as well as the regularity of the paths: they become smoother as H increases. Some of the basic properties of fractional Brownian motion are summarized in the following proposition. Proposition 1.27. Fractional Brownian motion has the following properties. 15 2 2 1.5 1.5 1 1 0.5 H t 0 H t 0.5 W W −0.5 0 −1 −0.5 −1.5 −2 −1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 t t a. H = 0.3 b. H = 0.8 Figure 1.4: Sample paths of fractional Brownian motion for Hurst exponent H = 0.3 and H = 0.8 and first (blue curve) and second (black curve) moment. 1 1 2 i. When H = 2 , Wt becomes the standard Brownian motion. ii. W H = 0, EW H = 0, E(W H)2 = t 2H , t > 0. 0 t t | | iii. It has stationary increments and E(W H W H )2 = t s 2H . t − s | − | iv. It has the following self similarity property H H H (Wαt , t > 0) = (α Wt , t > 0), α> 0, (1.28) where the equivalence is in law. The proof of these properties is left as an exercise. In Figure 1.4 we present sample plots and the first two moments of the factional Brownian motion for H = 0.3 and H = 0.8. As expected, for larger values of the Hurst exponent the sample paths are more regular. 1.5 The Karhunen-Loeve´ Expansion Let f L2( ) where is a subset of Rd and let e be an orthonormal basis in L2( ). Then, it is ∈ D D { n}n∞=1 D well known that f can be written as a series expansion: ∞ f = fnen, n=1 where fn = f(x)en(x) dx. Ω 16 The convergence is in L2( ): D N lim f(x) fnen(x) = 0. N − →∞ n=1 L2( ) D It turns out that we can obtain a similar expansion for an L2 mean zero process which is continuous in the L2 sense: E 2 E E 2 Xt < + , Xt = 0, lim Xt+h Xt = 0. (1.29) ∞ h 0 | − | → For simplicity we will take T = [0, 1]. Let R(t,s)= E(XtXs) be the autocorrelation function. Notice that from (1.29) it follows that R(t,s) is continuous in both t and s; see Exercise 20. Let us assume an expansion of the form ∞ X (ω)= ξ (ω)e (t), t [0, 1] (1.30) t n n ∈ n =1 where e is an orthonormal basis in L2(0, 1). The random variables ξ are calculated as { n}n∞=1 n 1 1 ∞ ∞ Xtek(t) dt = ξnen(t)ek(t) dt = ξnδnk = ξk, 0 0 n=1 n=1 where we assumed that we can interchange the summation and integration. We will assume that these random variables are orthogonal: E(ξnξm)= λnδnm, where λ are positive numbers that will be determined later. { n}n∞=1 Assuming that an expansion of the form (1.30) exists, we can calculate ∞ ∞ R(t,s)= E(XtXs) = E ξkek(t)ξℓeℓ(s) k=1 ℓ=1 ∞ ∞ = E (ξkξℓ) ek(t)eℓ(s) k=1 ℓ=1 ∞ = λkek(t)ek(s). k=1 Consequently, in order to the expansion (1.30) to be valid we need ∞ R(t,s)= λkek(t)ek(s). (1.31) k=1 17 From equation (1.31) it follows that 1 1 ∞ R(t,s)en(s) ds = λkek(t)ek(s)en(s) ds 0 0 k=1 ∞ 1 = λkek(t) ek(s)en(s) ds 0 k=1 ∞ = λkek(t)δkn k=1 = λnen(t). Consequently, in order for the expansion (1.30) to be valid, λ , e (t) have to be the eigenvalues and { n n }n∞=1 eigenfunctions of the integral operator whose kernel is the correlation function of Xt: 1 R(t,s)en(s) ds = λnen(t). (1.32) 0 To prove the expansion (1.30) we need to study the eigenvalue problem for the integral operator 1 f := R(t,s)f(s) ds. (1.33) R 0 We consider it as an operator from L2[0, 1] to L2[0, 1]. We can show that this operator is selfadjoint and nonnegative in L2(0, 1): f, h = f, h and f,f > 0 f, h L2(0, 1), R R R ∀ ∈ where , denotes the L2(0, 1)-inner product. It follows that all its eigenvalues are real and nonnegative. Furthermore, it is a compact operator (if φ is a bounded sequence in L2(0, 1), then φ has { n}n∞=1 {R n}n∞=1 a convergent subsequence). The spectral theorem for compact, selfadjoint operators can be used to deduce that has a countable sequence of eigenvalues tending to 0. Furthermore, for every f L2(0, 1) we can R ∈ write ∞ f = f0 + fnen(t), n=1 where f = 0 and e (t) are the eigenfunctions of the operator corresponding to nonzero eigenvalues R 0 { n } R and where the convergence is in L2. Finally, Mercer’s Theorem states that for R(t,s) continuous on [0, 1] × [0, 1], the expansion (1.31) is valid, where the series converges absolutely and uniformly. Now we are ready to prove (1.30). Theorem 1.28. (Karhunen-Loeve).´ Let X , t [0, 1] be an L2 process with zero mean and continuous { t ∈ } correlation function R(t,s). Let λ , e (t) be the eigenvalues and eigenfunctions of the operator { n n }n∞=1 R defined in (1.33). Then ∞ X = ξ e (t), t [0, 1], (1.34) t n n ∈ n =1 18 where 1 ξn = Xten(t) dt, Eξn = 0, E(ξnξm)= λδnm. (1.35) 0 The series converges in L2 to X(t), uniformly in t. Proof. The fact that Eξn = 0 follows from the fact that Xt is mean zero. The orthogonality of the random variables ξ follows from the orthogonality of the eigenfunctions of : { n}n∞=1 R 1 1 E(ξnξm) = E XtXsen(t)em(s) dtds 0 0 1 1 = R(t,s)en(t)em(s) dsdt 0 0 1 = λn en(s)em(s) ds = λnδnm. 0 N Consider now the partial sum SN = n=1 ξnen(t). E X S 2 = EX2 + ES2 2E(X S ) | t − N | t N − t N N N = R(t,t)+ E ξ ξ e (t)e (t) 2E X ξ e (t) k ℓ k ℓ − t n n n=1 k,ℓ =1 N N 1 2 = R(t,t)+ λk ek(t) 2E XtXsek(s)ek(t) ds | | − 0 k=1 k=1 N = R(t,t) λ e (t) 2 0, − k| k | → k=1 by Mercer’s theorem. The Karhunen-o´eve expansion is straightforward to apply to Gaussian stochastic processes. Let Xt be a Gaussian second order process with continuous covariance R(t,s). Then the random variables ξ are { k}k∞=1 Gaussian, since they are defined through the time integral of a Gaussian processes. Furthermore, since they are Gaussian and orthogonal, they are also independent. Hence, for Gaussian processes the Karhunen-Lo´eve expansion becomes: + ∞ Xt = λkξkek(t), (1.36) k=1 where ξ are independent (0, 1) random variables. { k}k∞=1 N Example 1.29. The Karhunen-Lo´eve Expansion for Brownian Motion. The correlation function of Brown- ian motion is R(t,s) = min(t,s). The eigenvalue problem ψ = λ ψ becomes R n n n 1 min(t,s)ψn(s) ds = λnψn(t). 0 19 Let us assume that λn > 0 (we can check that 0 is not an eigenvalue). Upon setting t = 0 we obtain ψn(0) = 0. The eigenvalue problem can be rewritten in the form t 1 sψn(s) ds + t ψn(s) ds = λnψn(t). 0 t We differentiate this equation once: 1 ψn(s) ds = λnψn′ (t). t We set t = 1 in this equation to obtain the second boundary condition ψn′ (1) = 0. A second differentiation yields; ψ (t)= λ ψ′′(t), − n n n where primes denote differentiation with respect to t. Thus, in order to calculate the eigenvalues and eigen- functions of the integral operator whose kernel is the covariance function of Brownian motion, we need to solve the Sturm-Liouville problem ψ (t)= λ ψ′′(t), ψ(0) = ψ′(1) = 0. − n n n We can calculate the eigenvalues and (normalized) eigenfunctions are 1 2 2 ψ (t)= √2sin (2n 1)πt , λ = . n 2 − n (2n 1)π − Thus, the Karhunen-Lo´eve expansion of Brownian motion on [0, 1] is ∞ 2 1 W = √2 ξ sin (2n 1)πt . (1.37) t n (2n 1)π 2 − n=1 − 1.6 Discussion and Bibliography The material presented in this chapter is very standard and can be found in any any textbook on stochastic processes. Consult, for example [48, 47, 49, 32]. The proof of Bochner’s theorem 1.14 can be found in [50], where additional material on stationary processes can be found. See also [48]. The Ornstein-Uhlenbeck process was introduced by Ornstein and Uhlenbeck in 1930 as a model for the velocity of a Brownian particle [101]. An early reference on the derivation of formulas of the form (1.15) is [99]. Gaussian processes are studied in [1]. Simulation algorithms for Gaussian processes are presented in [6]. Fractional Brownian motion was introduced in [64]. The spectral theorem for compact, selfadjoint operators that we used in the proof of the Karhunen-Lo´eve expansion can be found in [84]. The Karhunen-Lo´eve expansion can be used to generate random fields, i.e. a collection of random variables that are parameterized by a spatial (rather than temporal) parameter x. See [29]. The Karhunen-Lo´eve expansion is useful in the development of numerical algorithms for partial differential equations with random coefficients. See [92]. 20 We can use the Karhunen-Lo´eve expansion in order to study the L2-regularity of stochastic processes. First, let R be a compact, symmetric positive definite operator on L2(0, 1) with eigenvalues and normalized + 2 1 eigenfunctions λ , e (x) ∞ and consider a function f L (0, 1) with f(s) ds = 0. We can define { k k }k=1 ∈ 0 the one parameter family of Hilbert spaces Hα through the norm 2 α 2 2 α f = R− f 2 = f λ− . α L | k| k The inner product can be obtained through polarization. This norm enables us to measure the regularity 4 of the function f(t). Let Xt be a mean zero second order (i.e. with finite second moment) process with continuous autocorrelation function. Define the space α := L2((Ω, P ),Hα(0, 1)) with (semi)norm H 2 2 1 α X = E X α = λ − . (1.38) t α t H | k| k Notice that the regularity of the stochastic process Xt depends on the decay of the eigenvalues of the integral operator := 1 R(t,s) ds. R 0 As an example, consider the L2-regularity of Brownian motion. From Example 1.29 we know that λ k 2. Consequently, from (1.38) we get that, in order for W to be an element of the space α, we k ∼ − t H need that 2(1 α) λ − − < + , | k| ∞ k from which we obtain that α< 1/2. This is consistent with the H¨older continuity of Brownian motion from Theorem 1.22. 5 1.7 Exercises 1. Let Y0, Y1,... be a sequence of independent, identically distributed random variables and consider the stochastic process Xn = Yn. (a) Show that Xn is a strictly stationary process. (b) Assume that EY = < + and EY 2 = σ2 < + . Show that 0 ∞ 0 ∞ N 1 1 − lim E Xj = 0. N + N − → ∞ j=0 (c) Let f be such that Ef 2(Y ) < + . Show that 0 ∞ N 1 1 − lim E f(Xj) f(Y0) = 0. N + N − → ∞ j=0 4 α Think of R as being the inverse of the Laplacian with periodic boundary conditions. In this case H coincides with the standard fractional Sobolev space. 5Notice, however, that Wiener’s theorem refers to a.s. H¨older continuity, whereas the calculation presented in this section is about L2-continuity. 21 2. Let Z be a random variable and define the stochastic process Xn = Z, n = 0, 1, 2,... . Show that Xn is a strictly stationary process. 3. Let A0, A1,...Am and B0,B1,...Bm be uncorrelated random variables with mean zero and variances EA2 = σ2, EB2 = σ2, i = 1,...m. Let ω ,ω ,...ω [0,π] be distinct frequencies and define, for i i i i 0 1 m ∈ n = 0, 1, 2,... , the stochastic process ± ± m Xn = Ak cos(nωk)+ Bk sin(nωk) . k=0 Calculate the mean and the covariance of Xn. Show that it is a weakly stationary process. 4. Let ξ : n = 0, 1, 2,... be uncorrelated random variables with Eξ = , E(ξ )2 = σ2, n = { n ± ± } n n − 0, 1, 2,... . Let a , a ,... be arbitrary real numbers and consider the stochastic process ± ± 1 2 Xn = a1ξn + a2ξn 1 + . . . amξn m+1. − − (a) Calculate the mean, variance and the covariance function of Xn. Show that it is a weakly stationary process. (b) Set ak = 1/√m for k = 1,...m. Calculate the covariance function and study the cases m = 1 and m + . → ∞ 5. Let W (t) be a standard one dimensional Brownian motion. Calculate the following expectations. (a) EeiW (t). (b) Eei(W (t)+W (s)), t,s, (0, + ). ∈ ∞ (c) E( n c W (t ))2, where c R, i = 1,...n and t (0, + ), i = 1,...n. i=1 i i i ∈ i ∈ ∞ n i ciW (ti) (d) Ee Pi=1 , where c R, i = 1,...n and t (0, + ), i = 1,...n. i ∈ i ∈ ∞ 6. Let Wt be a standard one dimensional Brownian motion and define B = W tW , t [0, 1]. t t − 1 ∈ (a) Show that Bt is a Gaussian process with EB = 0, E(B B ) = min(t,s) ts. t t s − (b) Show that, for t [0, 1) an equivalent definition of B is through the formula ∈ t t B = (1 t)W . t − 1 t − (c) Calculate the distribution function of Bt. 22 7. Let Xt be a mean-zero second order stationary process with autocorrelation function N λ2 j αj t R(t)= e− | |, αj j=1 where α , λ N are positive real numbers. { j j}j=1 (a) Calculate the spectral density and the correlaction time of this process. (b) Show that the assumptions of Theorem 1.16 are satisfied and use the argument presented in Sec- tion 1.2 (i.e. the Green-Kubo formula) to calculate the diffusion coefficient of the process Zt = t 0 Xs ds. (c) Under what assumptions on the coefficients α , λ N can you study the above questions in the { j j}j=1 limit N + ? → ∞ 8. Show that the position of a Brownian particle whose velocity is described by the stationary Ornstein- Uhlenbeck process, Equation (1.10) is a mean zero Gaussian stochastic process and calculate the covari- ance function. 9. Let a1, . . . an and s1,...sn be positive real numbers. Calculate the mean and variance of the random variable n X = aiW (si). i=1 10. Let W (t) be the standard one-dimensional Brownian motion and let σ, s1, s2 > 0. Calculate (a) EeσW (t). (b) E sin(σW (s1)) sin(σW (s2)) . 11. Let Wt be a one dimensional Brownian motion and let ,σ > 0 and define t +σWt St = e . (a) Calculate the mean and the variance of St. (b) Calculate the probability density function of St. 12. Prove proposition 1.23. 13. Use Lemma 1.24 to calculate the distribution function of the stationary Ornstein-Uhlenbeck process. 14. Calculate the mean and the correlation function of the integral of a standard Brownian motion t Yt = Ws ds. 0 15. Show that the process t+1 Y = (W W ) ds, t R, t s − t ∈ t is second order stationary. 23 t 2t 16. Let Vt = e− W (e ) be the stationary Ornstein-Uhlenbeck process. Give the definition and study the main properties of the Ornstein-Uhlenbeck bridge. 17. The autocorrelation function of the velocity Y (t) a Brownian particle moving in a harmonic potential 1 2 2 V (x)= 2 ω0x is γ t 1 R(t)= e− | | cos(δ t ) sin(δ t ) , | | − δ | | where γ is the friction coefficient and δ = ω2 γ2. 0 − (a) Calculate the spectral density of Y (t). (b) Calculate the mean square displacement E(X(t))2 of the position of the Brownian particle X(t)= t Y (s) ds. Study the limit t + . 0 → ∞ 18. Show the scaling property (1.28) of the fractional Brownian motion. 19. The Poisson process with intensity λ, denoted by N(t), is an integer-valued, continuous time, stochastic process with independent increments satisfying λ(t s) k e− − λ(t s) P[(N(t) N(s)) = k]= − , t>s > 0, k N. − k! ∈ Use Theorem (1.21) to show that there does not exist a continuous modification of this process. 20. Show that the correlation function of a process Xt satisfying (1.29) is continuous in both t and s. 21. Let Xt be a stochastic process satisfying (1.29) and R(t,s) its correlation function. Show that the integral operator : L2[0, 1] L2[0, 1] defined in (1.33), R → 1 f := R(t,s)f(s) ds, R 0 is selfadjoint and nonnegative. Show that all of its eigenvalues are real and nonnegative. Show that eigenfunctions corresponding to different eigenvalues are orthogonal. 22. Let H be a Hilbert space. An operator : H H is said to be Hilbert–Schmidt if there exists a R → complete orthonormal sequence φ in H such that { n}n∞=1 ∞ e 2 < . R n ∞ n=1 Let : L2[0, 1] L2[0, 1] be the operator defined in (1.33) with R(t,s) being continuous both in t and R → s. Show that it is a Hilbert-Schmidt operator. 23. Let Xt a mean zero second order stationary process defined in the interval [0, T ] with continuous covari- + ance R(t) and let λ ∞ be the eigenvalues of the covariance operator. Show that { n}n=1 ∞ λn = TR(0). n=1 24 24. Calculate the Karhunen-Loeve expansion for a second order stochastic process with correlation function R(t,s)= ts. 25. Calculate the Karhunen-Loeve expansion of the Brownian bridge on [0, 1]. 26. Let X , t [0, T ] be a second order process with continuous covariance and Karhunen-Lo´eve expansion t ∈ ∞ Xt = ξkek(t). k=1 Define the process Y (t)= f(t)X , t [0,S], τ(t) ∈ where f(t) is a continuous function and τ(t) a continuous, nondecreasing function with τ(0) = 0, τ(S)= T . Find the Karhunen-Lo´eve expansion of Y (t), in an appropriate weighted L2 space, in terms of the KL expansion of Xt. Use this in order to calculate the KL expansion of the Ornstein-Uhlenbeck process. 27. Calculate the Karhunen-Lo´eve expansion of a centered Gaussian stochastic process with covariance func- tion R(s,t) = cos(2π(t s)). − 28. Use the Karhunen-Loeve expansion to generate paths of (a) the Brownian motion on [0, 1]; (b) the Brownian bridge on [0, 1]; (c) the Ornstein-Uhlenbeck on [0, 1]. Study computationally the convergence of the Karhunen-Lo´eve expansion for these processes. How many terms do you need to keep in the expansion in order to calculate accurate statistics of these pro- cesses? How does the computational cost compare with that of the standard algorithm for simulating Gaussian stochastic processes? 29. (See [29].) Consider the Gaussian random field X(x) in R with covariance function a x y γ(x,y)= e− | − | (1.39) where a> 0. (a) Simulate this field: generate samples and calculate the first four moments. (b) Consider X(x) for x [ L,L]. Calculate analytically the eigenvalues and eigenfunctions of the ∈ − integral operator with kernel γ(x,y), K L f(x)= γ(x,y)f(y) dy. K L − Use this in order to obtain the Karhunen-Lo´eve expansion for X. Plot the first five eigenfunctions when a = 1, L = 0.5. Investigate (either analytically or by means of numerical experiments) the − accuracy of the KL expansion as a function of the number of modes kept. 25 (c) Develop a numerical method for calculating the first few eigenvalues and eigenfunctions of with K a = 1, L = 0.5. Use the numerically calculated eigenvalues and eigenfunctions to simulate X(x) − using the KL expansion. Compare with the analytical results and comment on the accuracy of the calculation of the eigenvalues and eigenfunctions and on the computational cost. 26 Chapter 2 Diffusion Processes In this chapter we study some of the basic properties of Markov stochastic processes and, in particular, of diffusion processes. In Section 2.1 we present various examples of Markov processes, in discrete and continuous time. In Section 2.2 we give the precise definition of a Markov process. In Section 2.2 we derive the Chapman-Kolmogorov equation, the fundamental equation in the theory of Markov processes. In Section 2.3 we introduce the concept of the generator of a Markov process. In Section 2.4 we study ergodic Markov processes. In Section 2.5 we introduce diffusion processes and we derive the forward and backward Kolmogorov equations. Discussion and bibliographical remarks are presented in Section 2.6 and exercises can be found in Section 2.7. 2.1 Examples of Markov processes Roughly speaking, a Markov process is a stochastic process that retains no memory of where it has been in the past: only the current state of a Markov process can influence where it will go next. A bit more precisely: a Markov process is a stochastic process for which, given the present, the past and future are statistically independent. Perhaps the simplest example of a Markov process is that of a random walk in one dimension. Let ξi, i = 1,... be independent, identically distributed mean zero and variance 1 random variables. The one dimensional random walk is defined as N XN = ξn, X0 = 0. n=1 1 Let i1, i2,... be a sequence of integers. Then, for all integers n and m we have that P(X = i X = i ,...X = i )= P(X = i X = i ). (2.1) n+m n+m| 1 1 n n n+m n+m| n n In words, the probability that the random walk will be at in+m at time n + m depends only on its current value (at time n) and not on how it got there. 1In fact, it is sufficient to take m = 1 in (2.1). See Exercise 1. 27 The random walk is an example of a discrete time Markov chain: We will say that a stochastic process S ; n N with state space S = Z is a discrete time Markov chain provided that the Markov property (2.1) { n ∈ } is satisfied. Consider now a continuous-time stochastic process X with state space S = Z and denote by X , s 6 t { s t the collection of values of the stochastic process up to time t. We will say that X is a Markov processes } t provided that P(X = i X , s 6 t )= P(X = i X ), (2.2) t+h t+h|{ s } t+h t+h| t for all h > 0. A continuous-time, discrete state space Markov process is called a continuous-time Markov chain. A standard example of a continuous-time Markov chain is the Poisson process of rate λ with 0 if j < i, P(N = j N = i)= −λs j−i (2.3) t+h t e (λs) > | (j i)! , if j i. − Similarly, we can define a continuous-time Markov process with state space is R, as a stochastic process whose future depends on its present state and not on how it got there: P(X Γ X , s 6 t )= P(X Γ X ) (2.4) t+h ∈ |{ s } t+h ∈ | t for all Borel sets Γ. In this book we will consider continuous-time Markov processes for which a conditional probability density exists: P(X Γ X = x)= p(y,t + h x,t) dy. (2.5) t+h ∈ | t | Γ Example 2.1. The Brownian motion is a Markov process with conditional probability density given by the following formula 1 x y 2 P(Wt+h Γ Wt = x)= exp | − | dy. (2.6) ∈ | √2πh − 2h Γ The Markov property of Brownian motion follows from the fact that it has independent increments. t 2t Example 2.2. The stationary Ornstein-Uhlenbeck process Vt = e− W (e ) is a Markov process with con- ditional probability density 1 y xe (t s) 2 p(y,t x,s)= exp | − − − | . (2.7) 2(t s) 2(t s) | 2π(1 e ) −2(1 e− − ) − − − − To prove (2.7) we use the formula for the distribution function of the Brownian motion to calculate, for t>s, t 2t s 2s P(V 6 y V = x) = P(e− W (e ) 6 y e− W (e )= x) t | s | = P(W (e2t) 6 ety W (e2s)= esx) t | e y 1 |z−xes|2 = e− 2(e2t−e2s) dz 2π(e2t e2s) −∞ − y |ρet−xes|2 1 2t −2(t−s) = e− 2(e (1−e ) dρ 2πe2t(1 e 2(t s)) −∞ − − − y |ρ−x|2 1 −2(t−s) = e− 2(1−e ) dρ. 2π(1 e 2(t s)) −∞ − − − 28 Consequently, the transition probability density for the OU process is given by the formula ∂ p(y,t x,s) = P(V 6 y V = x) | ∂y t | s 1 y xe (t s) 2 = exp | − − − | . 2(t s) 2(t s) 2π(1 e ) −2(1 e− − ) − − − − The Markov property enables us to obtain an evolution equation for the transition probability for a discrete-time or continuous-time Markov chain P(X = i X = i ), P(X = i X = i ), (2.8) n+1 n+1| n n t+h t+h| t t or for the transition probability density defined in (2.5). This equation is the Chapman-Kolmogorov equa- tion. Using this equation we can study the evolution of a Markov process. We will be mostly concerned with time-homogeneous Markov processes, i.e. processes for which the conditional probabilities are invariant under time shifts. For time-homogeneous discrete-time Markov chains we have P(X = j X = i)= P(X = j X = i) =: p . n+1 | n 1 | 0 ij We will refer to the matrix P = p as the transition matrix. The transition matrix is a stochastic matrix, { ij} i.e. it has nonnegative entries and j pij = 1. Similarly, we can define the n-step transition matrix P = p (n) as n { ij } p (n)= P(X = j X = i). ij m+n | m We can study the evolution of a Markov chain through the Chapman-Kolmogorov equation: pij(m + n)= pik(m)pkj(n). (2.9) k (n) P (n) Indeed, let i := (Xn = i). The (possibly infinite dimensional) vector determines the state of the Markov chain at time n. From the Chapman-Kolmogorov equation we can obtain a formula for the evolution of the vector (n) (n) = (0)P n, (2.10) where P n denotes the nth power of the matrix P . Hence in order to calculate the state of the Markov chain at time n what we need is the initial distribution 0 and the transition matrix P . Componentwise, the above equation can be written as (n) (0) j = i πij(n). i Consider now a continuous-time Markov chain with transition probability p (s,t)= P(X = j X = i), s 6 t. ij t | s If the chain is homogeneous, then p (s,t)= p (0,t s) for all i, j, s, t. ij ij − 29 In particular, p (t)= P(X = j X = i). ij t | 0 The Chapman-Kolmogorov equation for a continuous-time Markov chain is dp ij = p (t)g , (2.11) dt ik kj k where the matrix G is called the generator of the Markov chain that is defined as 1 G = lim (Ph I), h 0 h − → with P denoting the matrix p (t) . Equation (2.11) can also be written in matrix form: t { ij } dP = P G. dt t i P Let now t = (Xt = i). The vector t is the distribution of the Markov chain at time t. We can study its evolution using the equation