Random Times and Their Properties
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Superprocesses and Mckean-Vlasov Equations with Creation of Mass
Sup erpro cesses and McKean-Vlasov equations with creation of mass L. Overb eck Department of Statistics, University of California, Berkeley, 367, Evans Hall Berkeley, CA 94720, y U.S.A. Abstract Weak solutions of McKean-Vlasov equations with creation of mass are given in terms of sup erpro cesses. The solutions can b e approxi- mated by a sequence of non-interacting sup erpro cesses or by the mean- eld of multityp e sup erpro cesses with mean- eld interaction. The lat- ter approximation is asso ciated with a propagation of chaos statement for weakly interacting multityp e sup erpro cesses. Running title: Sup erpro cesses and McKean-Vlasov equations . 1 Intro duction Sup erpro cesses are useful in solving nonlinear partial di erential equation of 1+ the typ e f = f , 2 0; 1], cf. [Dy]. Wenowchange the p oint of view and showhowtheyprovide sto chastic solutions of nonlinear partial di erential Supp orted byanFellowship of the Deutsche Forschungsgemeinschaft. y On leave from the Universitat Bonn, Institut fur Angewandte Mathematik, Wegelerstr. 6, 53115 Bonn, Germany. 1 equation of McKean-Vlasovtyp e, i.e. wewant to nd weak solutions of d d 2 X X @ @ @ + d x; + bx; : 1.1 = a x; t i t t t t t ij t @t @x @x @x i j i i=1 i;j =1 d Aweak solution = 2 C [0;T];MIR satis es s Z 2 t X X @ @ a f = f + f + d f + b f ds: s ij s t 0 i s s @x @x @x 0 i j i Equation 1.1 generalizes McKean-Vlasov equations of twodi erenttyp es. -
On Hitting Times for Jump-Diffusion Processes with Past Dependent Local Characteristics
Stochastic Processes and their Applications 47 ( 1993) 131-142 131 North-Holland On hitting times for jump-diffusion processes with past dependent local characteristics Manfred Schal Institut fur Angewandte Mathematik, Universitiit Bonn, Germany Received 30 March 1992 Revised 14 September 1992 It is well known how to apply a martingale argument to obtain the Laplace transform of the hitting time of zero (say) for certain processes starting at a positive level and being skip-free downwards. These processes have stationary and independent increments. In the present paper the method is extended to a more general class of processes the increments of which may depend both on time and past history. As a result a generalized Laplace transform is obtained which can be used to derive sharp bounds for the mean and the variance of the hitting time. The bounds also solve the control problem of how to minimize or maximize the expected time to reach zero. AMS 1980 Subject Classification: Primary 60G40; Secondary 60K30. hitting times * Laplace transform * expectation and variance * martingales * diffusions * compound Poisson process* random walks* M/G/1 queue* perturbation* control 1. Introduction Consider a process X,= S,- ct + uW, where Sis a compound Poisson process with Poisson parameter A which is disturbed by an independent standard Wiener process W This is the point of view of Dufresne and Gerber (1991). One can also look on X as a diffusion disturbed by jumps. This is the point of view of Ethier and Kurtz (1986, Section 4.10). Here u and A may be zero so that the jump term or the diffusion term may vanish. -
Stochastic Differential Equations
Stochastic Differential Equations Stefan Geiss November 25, 2020 2 Contents 1 Introduction 5 1.0.1 Wolfgang D¨oblin . .6 1.0.2 Kiyoshi It^o . .7 2 Stochastic processes in continuous time 9 2.1 Some definitions . .9 2.2 Two basic examples of stochastic processes . 15 2.3 Gaussian processes . 17 2.4 Brownian motion . 31 2.5 Stopping and optional times . 36 2.6 A short excursion to Markov processes . 41 3 Stochastic integration 43 3.1 Definition of the stochastic integral . 44 3.2 It^o'sformula . 64 3.3 Proof of Ito^'s formula in a simple case . 76 3.4 For extended reading . 80 3.4.1 Local time . 80 3.4.2 Three-dimensional Brownian motion is transient . 83 4 Stochastic differential equations 87 4.1 What is a stochastic differential equation? . 87 4.2 Strong Uniqueness of SDE's . 90 4.3 Existence of strong solutions of SDE's . 94 4.4 Theorems of L´evyand Girsanov . 98 4.5 Solutions of SDE's by a transformation of drift . 103 4.6 Weak solutions . 105 4.7 The Cox-Ingersoll-Ross SDE . 108 3 4 CONTENTS 4.8 The martingale representation theorem . 116 5 BSDEs 121 5.1 Introduction . 121 5.2 Setting . 122 5.3 A priori estimate . 123 Chapter 1 Introduction One goal of the lecture is to study stochastic differential equations (SDE's). So let us start with a (hopefully) motivating example: Assume that Xt is the share price of a company at time t ≥ 0 where we assume without loss of generality that X0 := 1. -
Notes on Elementary Martingale Theory 1 Conditional Expectations
. Notes on Elementary Martingale Theory by John B. Walsh 1 Conditional Expectations 1.1 Motivation Probability is a measure of ignorance. When new information decreases that ignorance, it changes our probabilities. Suppose we roll a pair of dice, but don't look immediately at the outcome. The result is there for anyone to see, but if we haven't yet looked, as far as we are concerned, the probability that a two (\snake eyes") is showing is the same as it was before we rolled the dice, 1/36. Now suppose that we happen to see that one of the two dice shows a one, but we do not see the second. We reason that we have rolled a two if|and only if|the unseen die is a one. This has probability 1/6. The extra information has changed the probability that our roll is a two from 1/36 to 1/6. It has given us a new probability, which we call a conditional probability. In general, if A and B are events, we say the conditional probability that B occurs given that A occurs is the conditional probability of B given A. This is given by the well-known formula P A B (1) P B A = f \ g; f j g P A f g providing P A > 0. (Just to keep ourselves out of trouble if we need to apply this to a set f g of probability zero, we make the convention that P B A = 0 if P A = 0.) Conditional probabilities are familiar, but that doesn't stop themf fromj g giving risef tog many of the most puzzling paradoxes in probability. -
The Theory of Filtrations of Subalgebras, Standardness and Independence
The theory of filtrations of subalgebras, standardness and independence Anatoly M. Vershik∗ 24.01.2017 In memory of my friend Kolya K. (1933{2014) \Independence is the best quality, the best word in all languages." J. Brodsky. From a personal letter. Abstract The survey is devoted to the combinatorial and metric theory of fil- trations, i. e., decreasing sequences of σ-algebras in measure spaces or decreasing sequences of subalgebras of certain algebras. One of the key notions, that of standardness, plays the role of a generalization of the no- tion of the independence of a sequence of random variables. We discuss the possibility of obtaining a classification of filtrations, their invariants, and various links to problems in algebra, dynamics, and combinatorics. Bibliography: 101 titles. Contents 1 Introduction 3 1.1 A simple example and a difficult question . .3 1.2 Three languages of measure theory . .5 1.3 Where do filtrations appear? . .6 1.4 Finite and infinite in classification problems; standardness as a generalization of independence . .9 arXiv:1705.06619v1 [math.DS] 18 May 2017 1.5 A summary of the paper . 12 2 The definition of filtrations in measure spaces 13 2.1 Measure theory: a Lebesgue space, basic facts . 13 2.2 Measurable partitions, filtrations . 14 2.3 Classification of measurable partitions . 16 ∗St. Petersburg Department of Steklov Mathematical Institute; St. Petersburg State Uni- versity Institute for Information Transmission Problems. E-mail: [email protected]. Par- tially supported by the Russian Science Foundation (grant No. 14-11-00581). 1 2.4 Classification of finite filtrations . 17 2.5 Filtrations we consider and how one can define them . -
THE MEASURABILITY of HITTING TIMES 1 Introduction
Elect. Comm. in Probab. 15 (2010), 99–105 ELECTRONIC COMMUNICATIONS in PROBABILITY THE MEASURABILITY OF HITTING TIMES RICHARD F. BASS1 Department of Mathematics, University of Connecticut, Storrs, CT 06269-3009 email: [email protected] Submitted January 18, 2010, accepted in final form March 8, 2010 AMS 2000 Subject classification: Primary: 60G07; Secondary: 60G05, 60G40 Keywords: Stopping time, hitting time, progressively measurable, optional, predictable, debut theorem, section theorem Abstract Under very general conditions the hitting time of a set by a stochastic process is a stopping time. We give a new simple proof of this fact. The section theorems for optional and predictable sets are easy corollaries of the proof. 1 Introduction A fundamental theorem in the foundations of stochastic processes is the one that says that, under very general conditions, the first time a stochastic process enters a set is a stopping time. The proof uses capacities, analytic sets, and Choquet’s capacibility theorem, and is considered hard. To the best of our knowledge, no more than a handful of books have an exposition that starts with the definition of capacity and proceeds to the hitting time theorem. (One that does is [1].) The purpose of this paper is to give a short and elementary proof of this theorem. The proof is simple enough that it could easily be included in a first year graduate course in probability. In Section 2 we give a proof of the debut theorem, from which the measurability theorem follows. As easy corollaries we obtain the section theorems for optional and predictable sets. This argument is given in Section 3. -
An Application of Probability Theory in Finance: the Black-Scholes Formula
AN APPLICATION OF PROBABILITY THEORY IN FINANCE: THE BLACK-SCHOLES FORMULA EFFY FANG Abstract. In this paper, we start from the building blocks of probability the- ory, including σ-field and measurable functions, and then proceed to a formal introduction of probability theory. Later, we introduce stochastic processes, martingales and Wiener processes in particular. Lastly, we present an appli- cation - the Black-Scholes Formula, a model used to price options in financial markets. Contents 1. The Building Blocks 1 2. Probability 4 3. Martingales 7 4. Wiener Processes 8 5. The Black-Scholes Formula 9 References 14 1. The Building Blocks Consider the set Ω of all possible outcomes of a random process. When the set is small, for example, the outcomes of tossing a coin, we can analyze case by case. However, when the set gets larger, we would need to study the collections of subsets of Ω that may be taken as a suitable set of events. We define such a collection as a σ-field. Definition 1.1. A σ-field on a set Ω is a collection F of subsets of Ω which obeys the following rules or axioms: (a) Ω 2 F (b) if A 2 F, then Ac 2 F 1 S1 (c) if fAngn=1 is a sequence in F, then n=1 An 2 F The pair (Ω; F) is called a measurable space. Proposition 1.2. Let (Ω; F) be a measurable space. Then (1) ; 2 F k SK (2) if fAngn=1 is a finite sequence of F-measurable sets, then n=1 An 2 F 1 T1 (3) if fAngn=1 is a sequence of F-measurable sets, then n=1 An 2 F Date: August 28,2015. -
Lecture 19 Semimartingales
Lecture 19:Semimartingales 1 of 10 Course: Theory of Probability II Term: Spring 2015 Instructor: Gordan Zitkovic Lecture 19 Semimartingales Continuous local martingales While tailor-made for the L2-theory of stochastic integration, martin- 2,c gales in M0 do not constitute a large enough class to be ultimately useful in stochastic analysis. It turns out that even the class of all mar- tingales is too small. When we restrict ourselves to processes with continuous paths, a naturally stable family turns out to be the class of so-called local martingales. Definition 19.1 (Continuous local martingales). A continuous adapted stochastic process fMtgt2[0,¥) is called a continuous local martingale if there exists a sequence ftngn2N of stopping times such that 1. t1 ≤ t2 ≤ . and tn ! ¥, a.s., and tn 2. fMt gt2[0,¥) is a uniformly integrable martingale for each n 2 N. In that case, the sequence ftngn2N is called the localizing sequence for (or is said to reduce) fMtgt2[0,¥). The set of all continuous local loc,c martingales M with M0 = 0 is denoted by M0 . Remark 19.2. 1. There is a nice theory of local martingales which are not neces- sarily continuous (RCLL), but, in these notes, we will focus solely on the continuous case. In particular, a “martingale” or a “local martingale” will always be assumed to be continuous. 2. While we will only consider local martingales with M0 = 0 in these notes, this is assumption is not standard, so we don’t put it into the definition of a local martingale. tn 3. -
Mathematical Finance an Introduction in Discrete Time
Jan Kallsen Mathematical Finance An introduction in discrete time CAU zu Kiel, WS 13/14, February 10, 2014 Contents 0 Some notions from stochastics 4 0.1 Basics . .4 0.1.1 Probability spaces . .4 0.1.2 Random variables . .5 0.1.3 Conditional probabilities and expectations . .6 0.2 Absolute continuity and equivalence . .7 0.3 Conditional expectation . .8 0.3.1 σ-fields . .8 0.3.2 Conditional expectation relative to a σ-field . 10 1 Discrete stochastic calculus 16 1.1 Stochastic processes . 16 1.2 Martingales . 18 1.3 Stochastic integral . 20 1.4 Intuitive summary of some notions from this chapter . 27 2 Modelling financial markets 36 2.1 Assets and trading strategies . 36 2.2 Arbitrage . 39 2.3 Dividend payments . 41 2.4 Concrete models . 43 3 Derivative pricing and hedging 48 3.1 Contingent claims . 49 3.2 Liquidly traded derivatives . 50 3.3 Individual perspective . 54 3.4 Examples . 56 3.4.1 Forward contract . 56 3.4.2 Futures contract . 57 3.4.3 European call and put options in the binomial model . 57 3.4.4 European call and put options in the standard model . 61 3.5 American options . 63 2 CONTENTS 3 4 Incomplete markets 70 4.1 Martingale modelling . 70 4.2 Variance-optimal hedging . 72 5 Portfolio optimization 73 5.1 Maximizing expected utility of terminal wealth . 73 6 Elements of continuous-time finance 80 6.1 Continuous-time theory . 80 6.2 Black-Scholes model . 84 Chapter 0 Some notions from stochastics 0.1 Basics 0.1.1 Probability spaces This course requires a decent background in stochastics which cannot be provided here. -
4 Martingales in Discrete-Time
4 Martingales in Discrete-Time Suppose that (Ω, F, P) is a probability space. Definition 4.1. A sequence F = {Fn, n = 0, 1,...} is called a filtration if each Fn is a sub-σ-algebra of F, and Fn ⊆ Fn+1 for n = 0, 1, 2,.... In other words, a filtration is an increasing sequence of sub-σ-algebras of F. For nota- tional convenience, we define ∞ ! [ F∞ = σ Fn n=1 and note that F∞ ⊆ F. Definition 4.2. If F = {Fn, n = 0, 1,...} is a filtration, and X = {Xn, n = 0, 1,...} is a discrete-time stochastic process, then X is said to be adapted to F if Xn is Fn-measurable for each n. Note that by definition every process {Xn, n = 0, 1,...} is adapted to its natural filtration {Fn, n = 0, 1,...} where Fn = σ(X0,...,Xn). The intuition behind the idea of a filtration is as follows. At time n, all of the information about ω ∈ Ω that is known to us at time n is contained in Fn. As n increases, our knowledge of ω also increases (or, if you prefer, does not decrease) and this is captured in the fact that the filtration consists of an increasing sequence of sub-σ-algebras. Definition 4.3. A discrete-time stochastic process X = {Xn, n = 0, 1,...} is called a martingale (with respect to the filtration F = {Fn, n = 0, 1,...}) if for each n = 0, 1, 2,..., (a) X is adapted to F; that is, Xn is Fn-measurable, 1 (b) Xn is in L ; that is, E|Xn| < ∞, and (c) E(Xn+1|Fn) = Xn a.s. -
Deep Optimal Stopping
Journal of Machine Learning Research 20 (2019) 1-25 Submitted 4/18; Revised 1/19; Published 4/19 Deep Optimal Stopping Sebastian Becker [email protected] Zenai AG, 8045 Zurich, Switzerland Patrick Cheridito [email protected] RiskLab, Department of Mathematics ETH Zurich, 8092 Zurich, Switzerland Arnulf Jentzen [email protected] SAM, Department of Mathematics ETH Zurich, 8092 Zurich, Switzerland Editor: Jon McAuliffe Abstract In this paper we develop a deep learning method for optimal stopping problems which directly learns the optimal stopping rule from Monte Carlo samples. As such, it is broadly applicable in situations where the underlying randomness can efficiently be simulated. We test the approach on three problems: the pricing of a Bermudan max-call option, the pricing of a callable multi barrier reverse convertible and the problem of optimally stopping a fractional Brownian motion. In all three cases it produces very accurate results in high- dimensional situations with short computing times. Keywords: optimal stopping, deep learning, Bermudan option, callable multi barrier reverse convertible, fractional Brownian motion 1. Introduction N We consider optimal stopping problems of the form supτ E g(τ; Xτ ), where X = (Xn)n=0 d is an R -valued discrete-time Markov process and the supremum is over all stopping times τ based on observations of X. Formally, this just covers situations where the stopping decision can only be made at finitely many times. But practically all relevant continuous- time stopping problems can be approximated with time-discretized versions. The Markov assumption means no loss of generality. We make it because it simplifies the presentation and many important problems already are in Markovian form. -
Optimal Stopping Time and Its Applications to Economic Models
OPTIMAL STOPPING TIME AND ITS APPLICATIONS TO ECONOMIC MODELS SIVAKORN SANGUANMOO Abstract. This paper gives an introduction to an optimal stopping problem by using the idea of an It^odiffusion. We then prove the existence and unique- ness theorems for optimal stopping, which will help us to explicitly solve opti- mal stopping problems. We then apply our optimal stopping theorems to the economic model of stock prices introduced by Samuelson [5] and the economic model of capital goods. Both economic models show that those related agents are risk-loving. Contents 1. Introduction 1 2. Preliminaries: It^odiffusion and boundary value problems 2 2.1. It^odiffusions and their properties 2 2.2. Harmonic function and the stochastic Dirichlet problem 4 3. Existence and uniqueness theorems for optimal stopping 5 4. Applications to an economic model: Stock prices 11 5. Applications to an economic model: Capital goods 14 Acknowledgments 18 References 19 1. Introduction Since there are several outside factors in the real world, many variables in classi- cal economic models are considered as stochastic processes. For example, the prices of stocks and oil do not just increase due to inflation over time but also fluctuate due to unpredictable situations. Optimal stopping problems are questions that ask when to stop stochastic pro- cesses in order to maximize utility. For example, when should one sell an asset in order to maximize his profit? When should one stop searching for the keyword to obtain the best result? Therefore, many recent economic models often include op- timal stopping problems. For example, by using optimal stopping, Choi and Smith [2] explored the effectiveness of the search engine, and Albrecht, Anderson, and Vroman [1] discovered how the search cost affects the search for job candidates.