Part II CONTINUOUS TIME STOCHASTIC PROCESSES
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
MAST31801 Mathematical Finance II, Spring 2019. Problem Sheet 8 (4-8.4.2019)
UH|Master program Mathematics and Statistics|MAST31801 Mathematical Finance II, Spring 2019. Problem sheet 8 (4-8.4.2019) 1. Show that a local martingale (Xt : t ≥ 0) in a filtration F, which is bounded from below is a supermartingale. Hint: use localization together with Fatou lemma. 2. A non-negative local martingale Xt ≥ 0 is a martingale if and only if E(Xt) = E(X0) 8t. 3. Recall Ito formula, for f(x; s) 2 C2;1 and a continuous semimartingale Xt with quadratic covariation hXit 1 Z t @2f Z t @f Z t @f f(Xt; t) − f(X0; 0) − 2 (Xs; s)dhXis − (Xs; s)ds = (Xs; s)dXs 2 0 @x 0 @s 0 @x where on the right side we have the Ito integral. We can also understand the Ito integral as limit in probability of Riemann sums X @f n n n n X(tk−1); tk−1 X(tk ^ t) − X(tk−1 ^ t) n n @x tk 2Π among a sequence of partitions Πn with ∆(Πn) ! 0. Recalling that the Brownian motion Wt has quadratic variation hW it = t, write the following Ito integrals explicitely by using Ito formula: R t n (a) 0 Ws dWs, n 2 N. R t (b) 0 exp σWs σdWs R t 2 (c) 0 exp σWs − σ s=2 σdWs R t (d) 0 sin(σWs)dWs R t (e) 0 cos(σWs)dWs 4. These Ito integrals as stochastic processes are semimartingales. Com- pute the quadratic variation of the Ito-integrals above. Hint: recall that the finite variation part of a semimartingale has zero quadratic variation, and the quadratic variation is bilinear. -
Poisson Processes Stochastic Processes
Poisson Processes Stochastic Processes UC3M Feb. 2012 Exponential random variables A random variable T has exponential distribution with rate λ > 0 if its probability density function can been written as −λt f (t) = λe 1(0;+1)(t) We summarize the above by T ∼ exp(λ): The cumulative distribution function of a exponential random variable is −λt F (t) = P(T ≤ t) = 1 − e 1(0;+1)(t) And the tail, expectation and variance are P(T > t) = e−λt ; E[T ] = λ−1; and Var(T ) = E[T ] = λ−2 The exponential random variable has the lack of memory property P(T > t + sjT > t) = P(T > s) Exponencial races In what follows, T1;:::; Tn are independent r.v., with Ti ∼ exp(λi ). P1: min(T1;:::; Tn) ∼ exp(λ1 + ··· + λn) . P2 λ1 P(T1 < T2) = λ1 + λ2 P3: λi P(Ti = min(T1;:::; Tn)) = λ1 + ··· + λn P4: If λi = λ and Sn = T1 + ··· + Tn ∼ Γ(n; λ). That is, Sn has probability density function (λs)n−1 f (s) = λe−λs 1 (s) Sn (n − 1)! (0;+1) The Poisson Process as a renewal process Let T1; T2;::: be a sequence of i.i.d. nonnegative r.v. (interarrival times). Define the arrival times Sn = T1 + ··· + Tn if n ≥ 1 and S0 = 0: The process N(t) = maxfn : Sn ≤ tg; is called Renewal Process. If the common distribution of the times is the exponential distribution with rate λ then process is called Poisson Process of with rate λ. Lemma. N(t) ∼ Poisson(λt) and N(t + s) − N(s); t ≥ 0; is a Poisson process independent of N(s); t ≥ 0 The Poisson Process as a L´evy Process A stochastic process fX (t); t ≥ 0g is a L´evyProcess if it verifies the following properties: 1. -
Perturbed Bessel Processes Séminaire De Probabilités (Strasbourg), Tome 32 (1998), P
SÉMINAIRE DE PROBABILITÉS (STRASBOURG) R.A. DONEY JONATHAN WARREN MARC YOR Perturbed Bessel processes Séminaire de probabilités (Strasbourg), tome 32 (1998), p. 237-249 <http://www.numdam.org/item?id=SPS_1998__32__237_0> © Springer-Verlag, Berlin Heidelberg New York, 1998, tous droits réservés. L’accès aux archives du séminaire de probabilités (Strasbourg) (http://portail. mathdoc.fr/SemProba/) implique l’accord avec les conditions générales d’utili- sation (http://www.numdam.org/conditions). Toute utilisation commerciale ou im- pression systématique est constitutive d’une infraction pénale. Toute copie ou im- pression de ce fichier doit contenir la présente mention de copyright. Article numérisé dans le cadre du programme Numérisation de documents anciens mathématiques http://www.numdam.org/ Perturbed Bessel Processes R.A.DONEY, J.WARREN, and M.YOR. There has been some interest in the literature in Brownian Inotion perturbed at its maximum; that is a process (Xt ; t > 0) satisfying (0.1) , where Mf = XS and (Bt; t > 0) is Brownian motion issuing from zero. The parameter a must satisfy a 1. For example arc-sine laws and Ray-Knight theorems have been obtained for this process; see Carmona, Petit and Yor [3], Werner [16], and Doney [7]. Our initial aim was to identify a process which could be considered as the process X conditioned to stay positive. This new process behaves like the Bessel process of dimension three except when at its maximum and we call it a perturbed three-dimensional Bessel process. We establish Ray-Knight theorems for the local times of this process, up to a first passage time and up to infinity (see Theorem 2.3), and observe that these descriptions coincide with those of the local times of two processes that have been considered in Yor [18]. -
A Stochastic Processes and Martingales
A Stochastic Processes and Martingales A.1 Stochastic Processes Let I be either IINorIR+.Astochastic process on I with state space E is a family of E-valued random variables X = {Xt : t ∈ I}. We only consider examples where E is a Polish space. Suppose for the moment that I =IR+. A stochastic process is called cadlag if its paths t → Xt are right-continuous (a.s.) and its left limits exist at all points. In this book we assume that every stochastic process is cadlag. We say a process is continuous if its paths are continuous. The above conditions are meant to hold with probability 1 and not to hold pathwise. A.2 Filtration and Stopping Times The information available at time t is expressed by a σ-subalgebra Ft ⊂F.An {F ∈ } increasing family of σ-algebras t : t I is called a filtration.IfI =IR+, F F F we call a filtration right-continuous if t+ := s>t s = t. If not stated otherwise, we assume that all filtrations in this book are right-continuous. In many books it is also assumed that the filtration is complete, i.e., F0 contains all IIP-null sets. We do not assume this here because we want to be able to change the measure in Chapter 4. Because the changed measure and IIP will be singular, it would not be possible to extend the new measure to the whole σ-algebra F. A stochastic process X is called Ft-adapted if Xt is Ft-measurable for all t. If it is clear which filtration is used, we just call the process adapted.The {F X } natural filtration t is the smallest right-continuous filtration such that X is adapted. -
Introduction to Stochastic Processes - Lecture Notes (With 33 Illustrations)
Introduction to Stochastic Processes - Lecture Notes (with 33 illustrations) Gordan Žitković Department of Mathematics The University of Texas at Austin Contents 1 Probability review 4 1.1 Random variables . 4 1.2 Countable sets . 5 1.3 Discrete random variables . 5 1.4 Expectation . 7 1.5 Events and probability . 8 1.6 Dependence and independence . 9 1.7 Conditional probability . 10 1.8 Examples . 12 2 Mathematica in 15 min 15 2.1 Basic Syntax . 15 2.2 Numerical Approximation . 16 2.3 Expression Manipulation . 16 2.4 Lists and Functions . 17 2.5 Linear Algebra . 19 2.6 Predefined Constants . 20 2.7 Calculus . 20 2.8 Solving Equations . 22 2.9 Graphics . 22 2.10 Probability Distributions and Simulation . 23 2.11 Help Commands . 24 2.12 Common Mistakes . 25 3 Stochastic Processes 26 3.1 The canonical probability space . 27 3.2 Constructing the Random Walk . 28 3.3 Simulation . 29 3.3.1 Random number generation . 29 3.3.2 Simulation of Random Variables . 30 3.4 Monte Carlo Integration . 33 4 The Simple Random Walk 35 4.1 Construction . 35 4.2 The maximum . 36 1 CONTENTS 5 Generating functions 40 5.1 Definition and first properties . 40 5.2 Convolution and moments . 42 5.3 Random sums and Wald’s identity . 44 6 Random walks - advanced methods 48 6.1 Stopping times . 48 6.2 Wald’s identity II . 50 6.3 The distribution of the first hitting time T1 .......................... 52 6.3.1 A recursive formula . 52 6.3.2 Generating-function approach . -
Patterns in Random Walks and Brownian Motion
Patterns in Random Walks and Brownian Motion Jim Pitman and Wenpin Tang Abstract We ask if it is possible to find some particular continuous paths of unit length in linear Brownian motion. Beginning with a discrete version of the problem, we derive the asymptotics of the expected waiting time for several interesting patterns. These suggest corresponding results on the existence/non-existence of continuous paths embedded in Brownian motion. With further effort we are able to prove some of these existence and non-existence results by various stochastic analysis arguments. A list of open problems is presented. AMS 2010 Mathematics Subject Classification: 60C05, 60G17, 60J65. 1 Introduction and Main Results We are interested in the question of embedding some continuous-time stochastic processes .Zu;0Ä u Ä 1/ into a Brownian path .BtI t 0/, without time-change or scaling, just by a random translation of origin in spacetime. More precisely, we ask the following: Question 1 Given some distribution of a process Z with continuous paths, does there exist a random time T such that .BTCu BT I 0 Ä u Ä 1/ has the same distribution as .Zu;0Ä u Ä 1/? The question of whether external randomization is allowed to construct such a random time T, is of no importance here. In fact, we can simply ignore Brownian J. Pitman ()•W.Tang Department of Statistics, University of California, 367 Evans Hall, Berkeley, CA 94720-3860, USA e-mail: [email protected]; [email protected] © Springer International Publishing Switzerland 2015 49 C. Donati-Martin et al. -
Derivatives of Self-Intersection Local Times
Derivatives of self-intersection local times Jay Rosen? Department of Mathematics College of Staten Island, CUNY Staten Island, NY 10314 e-mail: [email protected] Summary. We show that the renormalized self-intersection local time γt(x) for both the Brownian motion and symmetric stable process in R1 is differentiable in 0 the spatial variable and that γt(0) can be characterized as the continuous process of zero quadratic variation in the decomposition of a natural Dirichlet process. This Dirichlet process is the potential of a random Schwartz distribution. Analogous results for fractional derivatives of self-intersection local times in R1 and R2 are also discussed. 1 Introduction In their study of the intrinsic Brownian local time sheet and stochastic area integrals for Brownian motion, [14, 15, 16], Rogers and Walsh were led to analyze the functional Z t A(t, Bt) = 1[0,∞)(Bt − Bs) ds (1) 0 where Bt is a 1-dimensional Brownian motion. They showed that A(t, Bt) is not a semimartingale, and in fact showed that Z t Bs A(t, Bt) − Ls dBs (2) 0 x has finite non-zero 4/3-variation. Here Ls is the local time at x, which is x R s formally Ls = 0 δ(Br − x) dr, where δ(x) is Dirac’s ‘δ-function’. A formal d d2 0 application of Ito’s lemma, using dx 1[0,∞)(x) = δ(x) and dx2 1[0,∞)(x) = δ (x), yields Z t 1 Z t Z s Bs 0 A(t, Bt) − Ls dBs = t + δ (Bs − Br) dr ds (3) 0 2 0 0 ? This research was supported, in part, by grants from the National Science Foun- dation and PSC-CUNY. -
Skorohod Stochastic Integration with Respect to Non-Adapted Processes on Wiener Space Nicolas Privault
Skorohod stochastic integration with respect to non-adapted processes on Wiener space Nicolas Privault Equipe d'Analyse et Probabilit´es,Universit´ed'Evry-Val d'Essonne Boulevard des Coquibus, 91025 Evry Cedex, France e-mail: [email protected] Abstract We define a Skorohod type anticipative stochastic integral that extends the It^ointegral not only with respect to the Wiener process, but also with respect to a wide class of stochastic processes satisfying certain homogeneity and smoothness conditions, without requirements relative to filtrations such as adaptedness. Using this integral, a change of variable formula that extends the classical and Skorohod It^oformulas is obtained. Key words: It^ocalculus, Malliavin calculus, Skorohod integral. Mathematics Subject Classification (1991): 60H05, 60H07. 1 Introduction The Skorohod integral, defined by creation on Fock space, is an extension of the It^o integral with respect to the Wiener or Poisson processes, depending on the probabilis- tic interpretation chosen for the Fock space. This means that it coincides with the It^ointegral on square-integrable adapted processes. It can also extend the stochastic integral with respect to certain normal martingales, cf. [10], but it always acts with respect to the underlying process relative to a given probabilistic interpretation. The Skorohod integral is also linked to the Stratonovich, forward and backward integrals, and allows to extend the It^oformula to a class of Skorohod integral processes which contains non-adapted processes that have a certain structure. In this paper we introduce an extended Skorohod stochastic integral on Wiener space that can act with respect to processes other than Brownian motion. -
Lecture 19 Semimartingales
Lecture 19:Semimartingales 1 of 10 Course: Theory of Probability II Term: Spring 2015 Instructor: Gordan Zitkovic Lecture 19 Semimartingales Continuous local martingales While tailor-made for the L2-theory of stochastic integration, martin- 2,c gales in M0 do not constitute a large enough class to be ultimately useful in stochastic analysis. It turns out that even the class of all mar- tingales is too small. When we restrict ourselves to processes with continuous paths, a naturally stable family turns out to be the class of so-called local martingales. Definition 19.1 (Continuous local martingales). A continuous adapted stochastic process fMtgt2[0,¥) is called a continuous local martingale if there exists a sequence ftngn2N of stopping times such that 1. t1 ≤ t2 ≤ . and tn ! ¥, a.s., and tn 2. fMt gt2[0,¥) is a uniformly integrable martingale for each n 2 N. In that case, the sequence ftngn2N is called the localizing sequence for (or is said to reduce) fMtgt2[0,¥). The set of all continuous local loc,c martingales M with M0 = 0 is denoted by M0 . Remark 19.2. 1. There is a nice theory of local martingales which are not neces- sarily continuous (RCLL), but, in these notes, we will focus solely on the continuous case. In particular, a “martingale” or a “local martingale” will always be assumed to be continuous. 2. While we will only consider local martingales with M0 = 0 in these notes, this is assumption is not standard, so we don’t put it into the definition of a local martingale. tn 3. -
Local Time and Tanaka Formula for G-Brownian Motion
Local time and Tanaka formula for the G-Brownian motion∗ Qian LIN 1,2† 1School of Mathematics, Shandong University, Jinan 250100, China; 2 Laboratoire de Math´ematiques, CNRS UMR 6205, Universit´ede Bretagne Occidentale, 6, avenue Victor Le Gorgeu, CS 93837, 29238 Brest cedex 3, France. Abstract In this paper, we study the notion of local time and obtain the Tanaka formula for the G-Brownian motion. Moreover, the joint continuity of the local time of the G- Brownian motion is obtained and its quadratic variation is proven. As an application, we generalize Itˆo’s formula with respect to the G-Brownian motion to convex functions. Keywords: G-expectation; G-Brownian motion; local time; Tanaka formula; quadratic variation. 1 Introduction The objective of the present paper is to study the local time as well as the Tanaka formula for the G-Brownian motion. Motivated by uncertainty problems, risk measures and superhedging in finance, Peng has introduced a new notion of nonlinear expectation, the so-called G-expectation (see [10], [12], [13], [15], [16]), which is associated with the following nonlinear heat equation 2 ∂u(t,x) ∂ u(t,x) R arXiv:0912.1515v3 [math.PR] 20 Oct 2012 = G( 2 ), (t,x) [0, ) , ∂t ∂x ∈ ∞ × u(0,x)= ϕ(x), where, for given parameters 0 σ σ, the sublinear function G is defined as follows: ≤ ≤ 1 G(α)= (σ2α+ σ2α−), α R. 2 − ∈ The G-expectation represents a special case of general nonlinear expectations Eˆ whose importance stems from the fact that they are related to risk measures ρ in finance by ∗Corresponding address: Institute of Mathematical Economics, Bielefeld University, Postfach 100131, 33501 Bielefeld, Germany †Email address: [email protected] 1 the relation Eˆ[X] = ρ( X), where X runs the class of contingent claims. -
Non-Local Branching Superprocesses and Some Related Models
Published in: Acta Applicandae Mathematicae 74 (2002), 93–112. Non-local Branching Superprocesses and Some Related Models Donald A. Dawson1 School of Mathematics and Statistics, Carleton University, 1125 Colonel By Drive, Ottawa, Canada K1S 5B6 E-mail: [email protected] Luis G. Gorostiza2 Departamento de Matem´aticas, Centro de Investigaci´ony de Estudios Avanzados, A.P. 14-740, 07000 M´exicoD. F., M´exico E-mail: [email protected] Zenghu Li3 Department of Mathematics, Beijing Normal University, Beijing 100875, P.R. China E-mail: [email protected] Abstract A new formulation of non-local branching superprocesses is given from which we derive as special cases the rebirth, the multitype, the mass- structured, the multilevel and the age-reproduction-structured superpro- cesses and the superprocess-controlled immigration process. This unified treatment simplifies considerably the proof of existence of the old classes of superprocesses and also gives rise to some new ones. AMS Subject Classifications: 60G57, 60J80 Key words and phrases: superprocess, non-local branching, rebirth, mul- titype, mass-structured, multilevel, age-reproduction-structured, superprocess- controlled immigration. 1Supported by an NSERC Research Grant and a Max Planck Award. 2Supported by the CONACYT (Mexico, Grant No. 37130-E). 3Supported by the NNSF (China, Grant No. 10131040). 1 1 Introduction Measure-valued branching processes or superprocesses constitute a rich class of infinite dimensional processes currently under rapid development. Such processes arose in appli- cations as high density limits of branching particle systems; see e.g. Dawson (1992, 1993), Dynkin (1993, 1994), Watanabe (1968). The development of this subject has been stimu- lated from different subjects including branching processes, interacting particle systems, stochastic partial differential equations and non-linear partial differential equations. -
Simulation of Markov Chains
Copyright c 2007 by Karl Sigman 1 Simulating Markov chains Many stochastic processes used for the modeling of financial assets and other systems in engi- neering are Markovian, and this makes it relatively easy to simulate from them. Here we present a brief introduction to the simulation of Markov chains. Our emphasis is on discrete-state chains both in discrete and continuous time, but some examples with a general state space will be discussed too. 1.1 Definition of a Markov chain We shall assume that the state space S of our Markov chain is S = ZZ= f:::; −2; −1; 0; 1; 2;:::g, the integers, or a proper subset of the integers. Typical examples are S = IN = f0; 1; 2 :::g, the non-negative integers, or S = f0; 1; 2 : : : ; ag, or S = {−b; : : : ; 0; 1; 2 : : : ; ag for some integers a; b > 0, in which case the state space is finite. Definition 1.1 A stochastic process fXn : n ≥ 0g is called a Markov chain if for all times n ≥ 0 and all states i0; : : : ; i; j 2 S, P (Xn+1 = jjXn = i; Xn−1 = in−1;:::;X0 = i0) = P (Xn+1 = jjXn = i) (1) = Pij: Pij denotes the probability that the chain, whenever in state i, moves next (one unit of time later) into state j, and is referred to as a one-step transition probability. The square matrix P = (Pij); i; j 2 S; is called the one-step transition matrix, and since when leaving state i the chain must move to one of the states j 2 S, each row sums to one (e.g., forms a probability distribution): For each i X Pij = 1: j2S We are assuming that the transition probabilities do not depend on the time n, and so, in particular, using n = 0 in (1) yields Pij = P (X1 = jjX0 = i): (Formally we are considering only time homogenous MC's meaning that their transition prob- abilities are time-homogenous (time stationary).) The defining property (1) can be described in words as the future is independent of the past given the present state.