A Martingale Analysis of First Passage Times of Time-Dependent Wiener Diffusion Models
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Brownian Motion and the Heat Equation
Brownian motion and the heat equation Denis Bell University of North Florida 1. The heat equation Let the function u(t, x) denote the temperature in a rod at position x and time t u(t,x) Then u(t, x) satisfies the heat equation ∂u 1∂2u = , t > 0. (1) ∂t 2∂x2 It is easy to check that the Gaussian function 1 x2 u(t, x) = e−2t 2πt satisfies (1). Let φ be any! bounded continuous function and define 1 (x y)2 u(t, x) = ∞ φ(y)e− 2−t dy. 2πt "−∞ Then u satisfies!(1). Furthermore making the substitution z = (x y)/√t in the integral gives − 1 z2 u(t, x) = ∞ φ(x y√t)e−2 dz − 2π "−∞ ! 1 z2 φ(x) ∞ e−2 dz = φ(x) → 2π "−∞ ! as t 0. Thus ↓ 1 (x y)2 u(t, x) = ∞ φ(y)e− 2−t dy. 2πt "−∞ ! = E[φ(Xt)] where Xt is a N(x, t) random variable solves the heat equation ∂u 1∂2u = ∂t 2∂x2 with initial condition u(0, ) = φ. · Note: The function u(t, x) is smooth in x for t > 0 even if is only continuous. 2. Brownian motion In the nineteenth century, the botanist Robert Brown observed that a pollen particle suspended in liquid undergoes a strange erratic motion (caused by bombardment by molecules of the liquid) Letting w(t) denote the position of the particle in a fixed direction, the paths w typically look like this t N. Wiener constructed a rigorous mathemati- cal model of Brownian motion in the 1930s. -
1 Introduction Branching Mechanism in a Superprocess from a Branching
数理解析研究所講究録 1157 巻 2000 年 1-16 1 An example of random snakes by Le Gall and its applications 渡辺信三 Shinzo Watanabe, Kyoto University 1 Introduction The notion of random snakes has been introduced by Le Gall ([Le 1], [Le 2]) to construct a class of measure-valued branching processes, called superprocesses or continuous state branching processes ([Da], [Dy]). A main idea is to produce the branching mechanism in a superprocess from a branching tree embedded in excur- sions at each different level of a Brownian sample path. There is no clear notion of particles in a superprocess; it is something like a cloud or mist. Nevertheless, a random snake could provide us with a clear picture of historical or genealogical developments of”particles” in a superprocess. ” : In this note, we give a sample pathwise construction of a random snake in the case when the underlying Markov process is a Markov chain on a tree. A simplest case has been discussed in [War 1] and [Wat 2]. The construction can be reduced to this case locally and we need to consider a recurrence family of stochastic differential equations for reflecting Brownian motions with sticky boundaries. A special case has been already discussed by J. Warren [War 2] with an application to a coalescing stochastic flow of piece-wise linear transformations in connection with a non-white or non-Gaussian predictable noise in the sense of B. Tsirelson. 2 Brownian snakes Throughout this section, let $\xi=\{\xi(t), P_{x}\}$ be a Hunt Markov process on a locally compact separable metric space $S$ endowed with a metric $d_{S}(\cdot, *)$ . -
Arxiv:1909.00374V3 [Math.PR] 1 Apr 2021 2010 Mathematics Subject Classification
CONTRACTION PRINCIPLE FOR TRAJECTORIES OF RANDOM WALKS AND CRAMER'S´ THEOREM FOR KERNEL-WEIGHTED SUMS VLADISLAV VYSOTSKY Abstract. In 2013 A.A. Borovkov and A.A. Mogulskii proved a weaker-than-standard \metric" large deviations principle (LDP) for trajectories of random walks in Rd whose increments have the Laplace transform finite in a neighbourhood of zero. We prove that general metric LDPs are preserved under uniformly continuous mappings. This allows us to transform the result of Borovkov and Mogulskii into standard LDPs. We also give an explicit integral representation of the rate function they found. As an application, we extend the classical Cram´ertheorem by proving an LPD for kernel-weighted sums of i.i.d. random vectors in Rd. 1. Introduction The study of large deviations of trajectories of random walks was initiated by A.A. Borovkov in the 1960's. In 1976 A.A. Mogulskii [24] proved a large deviations result for the trajectories of a multidimensional random walk under the assumption that the Laplace transform of its increments is finite. In [24] Mogulskii also studied the large deviations under the weaker Cram´ermoment assumption, i.e. when the Laplace transform is finite in a neighbourhood of zero, but these results appear to be of a significantly limited use. The further progress was due to the concept of metric large deviations principles (LDPs, in short) on general metric spaces introduced by Borovkov and Mogulskii in [5] in 2010. The upper bound in such metric LDPs is worse than the conventional one { the infimum of the rate function is taken over shrinking "-neighbourhoods of a set rather than over its closure as in standard LDPs; compare definitions (1) and (2) below. -
Download File
i PREFACE Teaching stochastic processes to students whose primary interests are in applications has long been a problem. On one hand, the subject can quickly become highly technical and if mathe- matical concerns are allowed to dominate there may be no time available for exploring the many interesting areas of applications. On the other hand, the treatment of stochastic calculus in a cavalier fashion leaves the student with a feeling of great uncertainty when it comes to exploring new material. Moreover, the problem has become more acute as the power of the dierential equation point of view has become more widely appreciated. In these notes, an attempt is made to resolve this dilemma with the needs of those interested in building models and designing al- gorithms for estimation and control in mind. The approach is to start with Poisson counters and to identity the Wiener process with a certain limiting form. We do not attempt to dene the Wiener process per se. Instead, everything is done in terms of limits of jump processes. The Poisson counter and dierential equations whose right-hand sides include the dierential of Poisson counters are developed rst. This leads to the construction of a sample path representa- tions of a continuous time jump process using Poisson counters. This point of view leads to an ecient problem solving technique and permits a unied treatment of time varying and nonlinear problems. More importantly, it provides sound intuition for stochastic dierential equations and their uses without allowing the technicalities to dominate. In treating estimation theory, the conditional density equation is given a central role. -
Long Term Risk: a Martingale Approach
Long Term Risk: A Martingale Approach Likuan Qin∗ and Vadim Linetskyy Department of Industrial Engineering and Management Sciences McCormick School of Engineering and Applied Sciences Northwestern University Abstract This paper extends the long-term factorization of the pricing kernel due to Alvarez and Jermann (2005) in discrete time ergodic environments and Hansen and Scheinkman (2009) in continuous ergodic Markovian environments to general semimartingale environments, with- out assuming the Markov property. An explicit and easy to verify sufficient condition is given that guarantees convergence in Emery's semimartingale topology of the trading strate- gies that invest in T -maturity zero-coupon bonds to the long bond and convergence in total variation of T -maturity forward measures to the long forward measure. As applications, we explicitly construct long-term factorizations in generally non-Markovian Heath-Jarrow- Morton (1992) models evolving the forward curve in a suitable Hilbert space and in the non-Markovian model of social discount of rates of Brody and Hughston (2013). As a fur- ther application, we extend Hansen and Jagannathan (1991), Alvarez and Jermann (2005) and Bakshi and Chabi-Yo (2012) bounds to general semimartingale environments. When Markovian and ergodicity assumptions are added to our framework, we recover the long- term factorization of Hansen and Scheinkman (2009) and explicitly identify their distorted probability measure with the long forward measure. Finally, we give an economic interpre- tation of the recovery theorem of Ross (2013) in non-Markovian economies as a structural restriction on the pricing kernel leading to the growth optimality of the long bond and iden- tification of the physical measure with the long forward measure. -
Levy Processes
LÉVY PROCESSES, STABLE PROCESSES, AND SUBORDINATORS STEVEN P.LALLEY 1. DEFINITIONSAND EXAMPLES d Definition 1.1. A continuous–time process Xt = X(t ) t 0 with values in R (or, more generally, in an abelian topological groupG ) isf called a Lévyg ≥ process if (1) its sample paths are right-continuous and have left limits at every time point t , and (2) it has stationary, independent increments, that is: (a) For all 0 = t0 < t1 < < tk , the increments X(ti ) X(ti 1) are independent. − (b) For all 0 s t the··· random variables X(t ) X−(s ) and X(t s ) X(0) have the same distribution.≤ ≤ − − − The default initial condition is X0 = 0. A subordinator is a real-valued Lévy process with nondecreasing sample paths. A stable process is a real-valued Lévy process Xt t 0 with ≥ initial value X0 = 0 that satisfies the self-similarity property f g 1/α (1.1) Xt =t =D X1 t > 0. 8 The parameter α is called the exponent of the process. Example 1.1. The most fundamental Lévy processes are the Wiener process and the Poisson process. The Poisson process is a subordinator, but is not stable; the Wiener process is stable, with exponent α = 2. Any linear combination of independent Lévy processes is again a Lévy process, so, for instance, if the Wiener process Wt and the Poisson process Nt are independent then Wt Nt is a Lévy process. More important, linear combinations of independent Poisson− processes are Lévy processes: these are special cases of what are called compound Poisson processes: see sec. -
On the Form of the Large Deviation Rate Function for the Empirical Measures
Bernoulli 20(4), 2014, 1765–1801 DOI: 10.3150/13-BEJ540 On the form of the large deviation rate function for the empirical measures of weakly interacting systems MARKUS FISCHER Department of Mathematics, University of Padua, via Trieste 63, 35121 Padova, Italy. E-mail: fi[email protected] A basic result of large deviations theory is Sanov’s theorem, which states that the sequence of em- pirical measures of independent and identically distributed samples satisfies the large deviation principle with rate function given by relative entropy with respect to the common distribution. Large deviation principles for the empirical measures are also known to hold for broad classes of weakly interacting systems. When the interaction through the empirical measure corresponds to an absolutely continuous change of measure, the rate function can be expressed as relative entropy of a distribution with respect to the law of the McKean–Vlasov limit with measure- variable frozen at that distribution. We discuss situations, beyond that of tilted distributions, in which a large deviation principle holds with rate function in relative entropy form. Keywords: empirical measure; Laplace principle; large deviations; mean field interaction; particle system; relative entropy; Wiener measure 1. Introduction Weakly interacting systems are families of particle systems whose components, for each fixed number N of particles, are statistically indistinguishable and interact only through the empirical measure of the N-particle system. The study of weakly interacting systems originates in statistical mechanics and kinetic theory; in this context, they are often referred to as mean field systems. arXiv:1208.0472v4 [math.PR] 16 Oct 2014 The joint law of the random variables describing the states of the N-particle system of a weakly interacting system is invariant under permutations of components, hence determined by the distribution of the associated empirical measure. -
Large Deviation Theory
Large Deviation Theory J.M. Swart October 19, 2016 2 3 Preface The earliest origins of large deviation theory lie in the work of Boltzmann on en- tropy in the 1870ies and Cram´er'stheorem from 1938 [Cra38]. A unifying math- ematical formalism was only developed starting with Varadhan's definition of a `large deviation principle' (LDP) in 1966 [Var66]. Basically, large deviation theory centers around the observation that suitable func- tions F of large numbers of i.i.d. random variables (X1;:::;Xn) often have the property that −snI(x) P F (X1;:::;Xn) 2 dx ∼ e as n ! 1; (LDP) where sn are real contants such that limn!1 sn = 1 (in most cases simply sn = n). In words, (LDP) says that the probability that F (X1;:::;Xn) takes values near a point x decays exponentially fast, with speed sn, and rate function I. Large deviation theory has two different aspects. On the one hand, there is the question of how to formalize the intuitive formula (LDP). This leads to the al- ready mentioned definition of `large deviation principles' and involves quite a bit of measure theory and real analysis. The most important basic results of the ab- stract theory were proved more or less between 1966 and 1991, when O'Brian en Verwaat [OV91] and Puhalskii [Puk91] proved that exponential tightness implies a subsequential LDP. The abstract theory of large deviation principles plays more or less the same role as measure theory in (usual) probability theory. On the other hand, there is a much richer and much more important side of large deviation theory, which tries to identify rate functions I for various functions F of independent random variables, and study their properties. -
A Guide to Brownian Motion and Related Stochastic Processes
Vol. 0 (0000) A guide to Brownian motion and related stochastic processes Jim Pitman and Marc Yor Dept. Statistics, University of California, 367 Evans Hall # 3860, Berkeley, CA 94720-3860, USA e-mail: [email protected] Abstract: This is a guide to the mathematical theory of Brownian mo- tion and related stochastic processes, with indications of how this theory is related to other branches of mathematics, most notably the classical the- ory of partial differential equations associated with the Laplace and heat operators, and various generalizations thereof. As a typical reader, we have in mind a student, familiar with the basic concepts of probability based on measure theory, at the level of the graduate texts of Billingsley [43] and Durrett [106], and who wants a broader perspective on the theory of Brow- nian motion and related stochastic processes than can be found in these texts. Keywords and phrases: Markov process, random walk, martingale, Gaus- sian process, L´evy process, diffusion. AMS 2000 subject classifications: Primary 60J65. Contents 1 Introduction................................. 3 1.1 History ................................ 3 1.2 Definitions............................... 4 2 BM as a limit of random walks . 5 3 BMasaGaussianprocess......................... 7 3.1 Elementarytransformations . 8 3.2 Quadratic variation . 8 3.3 Paley-Wiener integrals . 8 3.4 Brownianbridges........................... 10 3.5 FinestructureofBrownianpaths . 10 arXiv:1802.09679v1 [math.PR] 27 Feb 2018 3.6 Generalizations . 10 3.6.1 FractionalBM ........................ 10 3.6.2 L´evy’s BM . 11 3.6.3 Browniansheets ....................... 11 3.7 References............................... 11 4 BMasaMarkovprocess.......................... 12 4.1 Markovprocessesandtheirsemigroups. 12 4.2 ThestrongMarkovproperty. 14 4.3 Generators ............................. -
Probability (Graduate Class) Lecture Notes
Probability (graduate class) Lecture Notes Tomasz Tkocz∗ These lecture notes were written for the graduate course 21-721 Probability that I taught at Carnegie Mellon University in Spring 2020. ∗Carnegie Mellon University; [email protected] 1 Contents 1 Probability space 6 1.1 Definitions . .6 1.2 Basic examples . .7 1.3 Conditioning . 11 1.4 Exercises . 13 2 Random variables 14 2.1 Definitions and basic properties . 14 2.2 π λ systems . 16 − 2.3 Properties of distribution functions . 16 2.4 Examples: discrete and continuous random variables . 18 2.5 Exercises . 21 3 Independence 22 3.1 Definitions . 22 3.2 Product measures and independent random variables . 24 3.3 Examples . 25 3.4 Borel-Cantelli lemmas . 28 3.5 Tail events and Kolmogorov's 0 1law .................. 29 − 3.6 Exercises . 31 4 Expectation 33 4.1 Definitions and basic properties . 33 4.2 Variance and covariance . 35 4.3 Independence again, via product measures . 36 4.4 Exercises . 39 5 More on random variables 41 5.1 Important distributions . 41 5.2 Gaussian vectors . 45 5.3 Sums of independent random variables . 46 5.4 Density . 47 5.5 Exercises . 49 6 Important inequalities and notions of convergence 52 6.1 Basic probabilistic inequalities . 52 6.2 Lp-spaces . 54 6.3 Notions of convergence . 59 6.4 Exercises . 63 2 7 Laws of large numbers 67 7.1 Weak law of large numbers . 67 7.2 Strong law of large numbers . 73 7.3 Exercises . 78 8 Weak convergence 81 8.1 Definition and equivalences . -
Stochastic Processes and Brownian Motion X(T) (Or X ) Is a Random Variable for Each Time T and • T Is Usually Called the State of the Process at Time T
Stochastic Processes A stochastic process • X = X(t) { } is a time series of random variables. Stochastic Processes and Brownian Motion X(t) (or X ) is a random variable for each time t and • t is usually called the state of the process at time t. A realization of X is called a sample path. • A sample path defines an ordinary function of t. • c 2006 Prof. Yuh-Dauh Lyuu, National Taiwan University Page 396 c 2006 Prof. Yuh-Dauh Lyuu, National Taiwan University Page 398 Stochastic Processes (concluded) Of all the intellectual hurdles which the human mind If the times t form a countable set, X is called a • has confronted and has overcome in the last discrete-time stochastic process or a time series. fifteen hundred years, the one which seems to me In this case, subscripts rather than parentheses are to have been the most amazing in character and • usually employed, as in the most stupendous in the scope of its consequences is the one relating to X = X . { n } the problem of motion. If the times form a continuum, X is called a — Herbert Butterfield (1900–1979) • continuous-time stochastic process. c 2006 Prof. Yuh-Dauh Lyuu, National Taiwan University Page 397 c 2006 Prof. Yuh-Dauh Lyuu, National Taiwan University Page 399 Random Walks The binomial model is a random walk in disguise. • Random Walk with Drift Consider a particle on the integer line, 0, 1, 2,... • ± ± Xn = µ + Xn−1 + ξn. In each time step, it can make one move to the right • with probability p or one move to the left with ξn are independent and identically distributed with zero • probability 1 p. -
Basics of the Theory of Large Deviations
Basics of the Theory of Large Deviations Michael Scheutzow February 1, 2018 Abstract In this note, we collect basic results of the theory of large deviations. Missing proofs can be found in the monograph [1]. 1 Introduction We start by recalling the following computation which was done in the course Wahrscheinlichkeitstheorie I (and which is also done in the course Versicherungs- mathematik). Assume that X; X1;X2; ::: are i.i.d real-valued random variables on a proba- bility space (Ω; F; P) and x 2 R, λ > 0. Then, by Markov’s inequality, n n X λX1 P Xi ≥ nx ≤ exp{−λnxg E e = exp{−n(λx − Λ(λ))g; i=1 where Λ(λ) := log E expfλXg. Defining I(x) := supλ≥0fλx − Λ(λ)g, we therefore get n X P Xi ≥ nx ≤ exp{−nI(x)g; i=1 which often turns out to be a good bound. 1 2 Cramer’s theorem for real-valued random vari- ables Definition 2.1. a) Let X be a real-valued random variable. The function Λ: R ! (−∞; 1] defined by λX Λ(λ) := log Ee is called cumulant generating function or logarithmic moment generating function. Let DΛ := fλ : Λ(λ) < 1g. b) Λ∗ : R ! [0; 1] defined by Λ∗(x) := supfλx − Λ(λ)g λ2R ∗ is called Fenchel-Legendre transform of Λ. Let DΛ∗ := fλ :Λ (λ) < 1g. In the following we will often use the convenient abbreviation Λ∗(F ) := inf Λ∗(x) x2F for a subset F ⊆ R (with the usual convention that the infimum of the empty set is +1).