Operator Methods for Continuous-Time Markov Processes
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Persistent Random Walks. II. Functional Scaling Limits
Persistent random walks. II. Functional Scaling Limits Peggy Cénac1. Arnaud Le Ny2. Basile de Loynes3. Yoann Offret1 1Institut de Mathématiques de Bourgogne (IMB) - UMR CNRS 5584 Université de Bourgogne Franche-Comté, 21000 Dijon, France 2Laboratoire d’Analyse et de Mathématiques Appliquées (LAMA) - UMR CNRS 8050 Université Paris Est Créteil, 94010 Créteil Cedex, France 3Ensai - Université de Bretagne-Loire, Campus de Ker-Lann, Rue Blaise Pascal, BP 37203, 35172 BRUZ cedex, France Abstract We give a complete and unified description – under some stability assumptions – of the functional scaling limits associated with some persistent random walks for which the recurrent or transient type is studied in [1]. As a result, we highlight a phase transition phenomenon with respect to the memory. It turns out that the limit process is either Markovian or not according to – to put it in a nutshell – the rate of decrease of the distribution tails corresponding to the persistent times. In the memoryless situation, the limits are classical strictly stable Lévy processes of infinite variations. However, we point out that the description of the critical Cauchy case fills some lacuna even in the closely related context of Directionally Reinforced Random Walks (DRRWs) for which it has not been considered yet. Besides, we need to introduced some relevant generalized drift – extended the classical one – in order to study the critical case but also the situation when the limit is no longer Markovian. It appears to be in full generality a drift in mean for the Persistent Random Walk (PRW). The limit processes keeping some memory – given by some variable length Markov chain – of the underlying PRW are called arcsine Lamperti anomalous diffusions due to their marginal distribution which are computed explicitly here. -
Math 353 Lecture Notes Intro to Pdes Eigenfunction Expansions for Ibvps
Math 353 Lecture Notes Intro to PDEs Eigenfunction expansions for IBVPs J. Wong (Fall 2020) Topics covered • Eigenfunction expansions for PDEs ◦ The procedure for time-dependent problems ◦ Projection, independent evolution of modes 1 The eigenfunction method to solve PDEs We are now ready to demonstrate how to use the components derived thus far to solve the heat equation. First, two examples to illustrate the proces... 1.1 Example 1: no source; Dirichlet BCs The simplest case. We solve ut =uxx; x 2 (0; 1); t > 0 (1.1a) u(0; t) = 0; u(1; t) = 0; (1.1b) u(x; 0) = f(x): (1.1c) The eigenvalues/eigenfunctions are (as calculated in previous sections) 2 2 λn = n π ; φn = sin nπx; n ≥ 1: (1.2) Assuming the solution exists, it can be written in the eigenfunction basis as 1 X u(x; t) = cn(t)φn(x): n=0 Definition (modes) The n-th term of this series is sometimes called the n-th mode or Fourier mode. I'll use the word frequently to describe it (rather than, say, `basis function'). 1 00 Substitute into the PDE (1.1a) and use the fact that −φn = λnφ to obtain 1 X 0 (cn(t) + λncn(t))φn(x) = 0: n=1 By the fact the fφng is a basis, it follows that the coefficient for each mode satisfies the ODE 0 cn(t) + λncn(t) = 0: Solving the ODE gives us a `general' solution to the PDE with its BCs, 1 X −λnt u(x; t) = ane φn(x): n=1 The remaining coefficients are determined by the IC, u(x; 0) = f(x): To match to the solution, we need to also write f(x) in the basis: 1 Z 1 X hf; φni f(x) = f φ (x); f = = 2 f(x) sin nπx dx: (1.3) n n n hφ ; φ i n=1 n n 0 Then from the initial condition, we get u(x; 0) = f(x) 1 1 X X =) cn(0)φn(x) = fnφn(x) n=1 n=1 =) cn(0) = fn for all n ≥ 1: Now everything has been solved - we are done! The solution to the IBVP (1.1) is 1 X −n2π2t u(x; t) = ane sin nπx with an given by (1.5): (1.4) n=1 Alternatively, we could state the solution as follows: The solution is 1 X −λnt u(x; t) = fne φn(x) n=1 with eigenfunctions/values φn; λn given by (1.2) and fn by (1.3). -
Math 356 Lecture Notes Intro to Pdes Eigenfunction Expansions for Ibvps
MATH 356 LECTURE NOTES INTRO TO PDES EIGENFUNCTION EXPANSIONS FOR IBVPS J. WONG (FALL 2019) Topics covered • Eigenfunction expansions for PDEs ◦ The procedure for time-dependent problems ◦ Projection, independent evolution of modes ◦ Separation of variables • The wave equation ◦ Standard solution, standing waves ◦ Example: tuning the vibrating string (resonance) • Interpreting solutions (what does the series mean?) ◦ Relevance of Fourier convergence theorem ◦ Smoothing for the heat equation ◦ Non-smoothing for the wave equation 3 1. The eigenfunction method to solve PDEs We are now ready to demonstrate how to use the components derived thus far to solve the heat equation. The procedure here for the heat equation will extend nicely to a variety of other problems. For now, consider an initial boundary value problem of the form ut = −Lu + h(x; t); x 2 (a; b); t > 0 hom. BCs at a and b (1.1) u(x; 0) = f(x) We seek a solution in terms of the eigenfunction basis X u(x; t) = cn(t)φn(x) n by finding simple ODEs to solve for the coefficients cn(t): This form of the solution is called an eigenfunction expansion for u (or `eigenfunction series') and each term cnφn(x) is a mode (or `Fourier mode' or `eigenmode'). Part 1: find the eigenfunction basis. The first step is to compute the basis. The eigenfunctions we need are the solutions to the eigenvalue problem Lφ = λφ, φ(x) satisfies the BCs for u: (1.2) By the theorem in ??, there is a sequence of eigenfunctions fφng with eigenvalues fλng that form an orthogonal basis for L2[a; b] (i.e. -
Lectures on Lévy Processes, Stochastic Calculus and Financial
Lectures on L¶evyProcesses, Stochastic Calculus and Financial Applications, Ovronnaz September 2005 David Applebaum Probability and Statistics Department, University of She±eld, Hicks Building, Houns¯eld Road, She±eld, England, S3 7RH e-mail: D.Applebaum@she±eld.ac.uk Introduction A L¶evyprocess is essentially a stochastic process with stationary and in- dependent increments. The basic theory was developed, principally by Paul L¶evyin the 1930s. In the past 15 years there has been a renaissance of interest and a plethora of books, articles and conferences. Why ? There are both theoretical and practical reasons. Theoretical ² There are many interesting examples - Brownian motion, simple and compound Poisson processes, ®-stable processes, subordinated processes, ¯nancial processes, relativistic process, Riemann zeta process . ² L¶evyprocesses are simplest generic class of process which have (a.s.) continuous paths interspersed with random jumps of arbitrary size oc- curring at random times. ² L¶evyprocesses comprise a natural subclass of semimartingales and of Markov processes of Feller type. ² Noise. L¶evyprocesses are a good model of \noise" in random dynamical systems. 1 Input + Noise = Output Attempts to describe this di®erentially leads to stochastic calculus.A large class of Markov processes can be built as solutions of stochastic di®erential equations driven by L¶evynoise. L¶evydriven stochastic partial di®erential equations are beginning to be studied with some intensity. ² Robust structure. Most applications utilise L¶evyprocesses taking val- ues in Euclidean space but this can be replaced by a Hilbert space, a Banach space (these are important for spdes), a locally compact group, a manifold. Quantised versions are non-commutative L¶evyprocesses on quantum groups. -
A Stochastic Processes and Martingales
A Stochastic Processes and Martingales A.1 Stochastic Processes Let I be either IINorIR+.Astochastic process on I with state space E is a family of E-valued random variables X = {Xt : t ∈ I}. We only consider examples where E is a Polish space. Suppose for the moment that I =IR+. A stochastic process is called cadlag if its paths t → Xt are right-continuous (a.s.) and its left limits exist at all points. In this book we assume that every stochastic process is cadlag. We say a process is continuous if its paths are continuous. The above conditions are meant to hold with probability 1 and not to hold pathwise. A.2 Filtration and Stopping Times The information available at time t is expressed by a σ-subalgebra Ft ⊂F.An {F ∈ } increasing family of σ-algebras t : t I is called a filtration.IfI =IR+, F F F we call a filtration right-continuous if t+ := s>t s = t. If not stated otherwise, we assume that all filtrations in this book are right-continuous. In many books it is also assumed that the filtration is complete, i.e., F0 contains all IIP-null sets. We do not assume this here because we want to be able to change the measure in Chapter 4. Because the changed measure and IIP will be singular, it would not be possible to extend the new measure to the whole σ-algebra F. A stochastic process X is called Ft-adapted if Xt is Ft-measurable for all t. If it is clear which filtration is used, we just call the process adapted.The {F X } natural filtration t is the smallest right-continuous filtration such that X is adapted. -
Lectures on Multiparameter Processes
Lecture Notes on Multiparameter Processes: Ecole Polytechnique Fed´ erale´ de Lausanne, Switzerland Davar Khoshnevisan Department of Mathematics University of Utah Salt Lake City, UT 84112–0090 [email protected] http://www.math.utah.edu/˜davar April–June 2001 ii Contents Preface vi 1 Examples from Markov chains 1 2 Examples from Percolation on Trees and Brownian Motion 7 3ProvingLevy’s´ Theorem and Introducing Martingales 13 4 Preliminaries on Ortho-Martingales 19 5 Ortho-Martingales and Intersections of Walks and Brownian Motion 25 6 Intersections of Brownian Motion, Multiparameter Martingales 35 7 Capacity, Energy and Dimension 43 8 Frostman’s Theorem, Hausdorff Dimension and Brownian Motion 49 9 Potential Theory of Brownian Motion and Stable Processes 55 10 Brownian Sheet and Kahane’s Problem 65 Bibliography 71 iii iv Preface These are the notes for a one-semester course based on ten lectures given at the Ecole Polytechnique Fed´ erale´ de Lausanne, April–June 2001. My goal has been to illustrate, in some detail, some of the salient features of the theory of multiparameter processes and in particular, Cairoli’s theory of multiparameter mar- tingales. In order to get to the heart of the matter, and develop a kind of intuition at the same time, I have chosen the simplest topics of random walks, Brownian motions, etc. to highlight the methods. The full theory can be found in Multi-Parameter Processes: An Introduction to Random Fields (henceforth, referred to as MPP) which is to be published by Springer-Verlag, although these lectures also contain material not covered in the mentioned book. -
Introduction to Lévy Processes
Introduction to L´evyprocesses Graduate lecture 22 January 2004 Matthias Winkel Departmental lecturer (Institute of Actuaries and Aon lecturer in Statistics) 1. Random walks and continuous-time limits 2. Examples 3. Classification and construction of L´evy processes 4. Examples 5. Poisson point processes and simulation 1 1. Random walks and continuous-time limits 4 Definition 1 Let Yk, k ≥ 1, be i.i.d. Then n X 0 Sn = Yk, n ∈ N, k=1 is called a random walk. -4 0 8 16 Random walks have stationary and independent increments Yk = Sk − Sk−1, k ≥ 1. Stationarity means the Yk have identical distribution. Definition 2 A right-continuous process Xt, t ∈ R+, with stationary independent increments is called L´evy process. 2 Page 1 What are Sn, n ≥ 0, and Xt, t ≥ 0? Stochastic processes; mathematical objects, well-defined, with many nice properties that can be studied. If you don’t like this, think of a model for a stock price evolving with time. There are also many other applications. If you worry about negative values, think of log’s of prices. What does Definition 2 mean? Increments , = 1 , are independent and Xtk − Xtk−1 k , . , n , = 1 for all 0 = . Xtk − Xtk−1 ∼ Xtk−tk−1 k , . , n t0 < . < tn Right-continuity refers to the sample paths (realisations). 3 Can we obtain L´evyprocesses from random walks? What happens e.g. if we let the time unit tend to zero, i.e. take a more and more remote look at our random walk? If we focus at a fixed time, 1 say, and speed up the process so as to make n steps per time unit, we know what happens, the answer is given by the Central Limit Theorem: 2 Theorem 1 (Lindeberg-L´evy) If σ = V ar(Y1) < ∞, then Sn − (Sn) √E → Z ∼ N(0, σ2) in distribution, as n → ∞. -
6 Mar 2019 Strict Local Martingales and the Khasminskii Test for Explosions
Strict Local Martingales and the Khasminskii Test for Explosions Aditi Dandapani∗† and Philip Protter‡§ March 7, 2019 Abstract We exhibit sufficient conditions such that components of a multidimen- sional SDE giving rise to a local martingale M are strict local martingales or martingales. We assume that the equations have diffusion coefficients of the form σ(Mt, vt), with vt being a stochastic volatility term. We dedicate this paper to the memory of Larry Shepp 1 Introduction In 2007 P.L. Lions and M. Musiela (2007) gave a sufficient condition for when a unique weak solution of a one dimensional stochastic volatility stochastic differen- tial equation is a strict local martingale. Lions and Musiela also gave a sufficient condition for it to be a martingale. Similar results were obtained by L. Andersen and V. Piterbarg (2007). The techniques depend at one key stage on Feller’s test for explosions, and as such are uniquely applicable to the one dimensional case. Earlier, in 2002, F. Delbaen and H. Shirakawa (2002) gave a seminal result giving arXiv:1903.02383v1 [q-fin.MF] 6 Mar 2019 necessary and sufficient conditions for a solution of a one dimensional stochastic differential equation (without stochastic volatility) to be a strict local martingale or not. This was later extended by A. Mijatovic and M. Urusov (2012) and then by ∗Applied Mathematics Department, Columbia University, New York, NY 10027; email: [email protected]; Currently at Ecole Polytechnique, Palaiseau, France. †Supported in part by NSF grant DMS-1308483 ‡Statistics Department, Columbia University, New York, NY 10027; email: [email protected]. -
Sturm-Liouville Expansions of the Delta
Gauge Institute Journal H. Vic Dannon Sturm-Liouville Expansions of the Delta Function H. Vic Dannon [email protected] May, 2014 Abstract We expand the Delta Function in Series, and Integrals of Sturm-Liouville Eigen-functions. Keywords: Sturm-Liouville Expansions, Infinitesimal, Infinite- Hyper-Real, Hyper-Real, infinite Hyper-real, Infinitesimal Calculus, Delta Function, Fourier Series, Laguerre Polynomials, Legendre Functions, Bessel Functions, Delta Function, 2000 Mathematics Subject Classification 26E35; 26E30; 26E15; 26E20; 26A06; 26A12; 03E10; 03E55; 03E17; 03H15; 46S20; 97I40; 97I30. 1 Gauge Institute Journal H. Vic Dannon Contents 0. Eigen-functions Expansion of the Delta Function 1. Hyper-real line. 2. Hyper-real Function 3. Integral of a Hyper-real Function 4. Delta Function 5. Convergent Series 6. Hyper-real Sturm-Liouville Problem 7. Delta Expansion in Non-Normalized Eigen-Functions 8. Fourier-Sine Expansion of Delta Associated with yx"( )+=l yx ( ) 0 & ab==0 9. Fourier-Cosine Expansion of Delta Associated with 1 yx"( )+=l yx ( ) 0 & ab==2 p 10. Fourier-Sine Expansion of Delta Associated with yx"( )+=l yx ( ) 0 & a = 0 11. Fourier-Bessel Expansion of Delta Associated with n2 -1 ux"( )+- (l 4 ) ux ( ) = 0 & ab==0 x 2 12. Delta Expansion in Orthonormal Eigen-functions 13. Fourier-Hermit Expansion of Delta Associated with ux"( )+- (l xux2 ) ( ) = 0 for any real x 14. Fourier-Legendre Expansion of Delta Associated with 2 Gauge Institute Journal H. Vic Dannon 11 11 uu"(ql )++ [ ] ( q ) = 0 -<<pq p 4 cos2 q , 22 15. Fourier-Legendre Expansion of Delta Associated with 112 11 ymy"(ql )++- [ ( ) ] (q ) = 0 -<<pq p 4 cos2 q , 22 16. -
Lecture 15.Pdf
Lecture #15 Eigenfunctions of hermitian operators If the spectrum is discrete , then the eigenfunctions are in Hilbert space and correspond to physically realizable states. It the spectrum is continuous , the eigenfunctions are not normalizable, and they do not correspond to physically realizable states (but their linear combinations may be normalizable). Discrete spectra Normalizable eigenfunctions of a hermitian operator have the following properties: (1) Their eigenvalues are real. (2) Eigenfunctions that belong to different eigenvalues are orthogonal. Finite-dimension vector space: The eigenfunctions of an observable operator are complete , i.e. any function in Hilbert space can be expressed as their linear combination. Continuous spectra Example: Find the eigenvalues and eigenfunctions of the momentum operator . This solution is not square -inferable and operator p has no eigenfunctions in Hilbert space. However, if we only consider real eigenvalues, we can define sort of orthonormality: Lecture 15 Page 1 L15.P2 since the Fourier transform of Dirac delta function is Proof: Plancherel's theorem Now, Then, which looks very similar to orthonormality. We can call such equation Dirac orthonormality. These functions are complete in a sense that any square integrable function can be written in a form and c(p) is obtained by Fourier's trick. Summary for continuous spectra: eigenfunctions with real eigenvalues are Dirac orthonormalizable and complete. Lecture 15 Page 2 L15.P3 Generalized statistical interpretation: If your measure observable Q on a particle in a state you will get one of the eigenvalues of the hermitian operator Q. If the spectrum of Q is discrete, the probability of getting the eigenvalue associated with orthonormalized eigenfunction is It the spectrum is continuous, with real eigenvalues q(z) and associated Dirac-orthonormalized eigenfunctions , the probability of getting a result in the range dz is The wave function "collapses" to the corresponding eigenstate upon measurement. -
Part C Lévy Processes and Finance
Part C Levy´ Processes and Finance Matthias Winkel1 University of Oxford HT 2007 1Departmental lecturer (Institute of Actuaries and Aon Lecturer in Statistics) at the Department of Statistics, University of Oxford MS3 Levy´ Processes and Finance Matthias Winkel – 16 lectures HT 2007 Prerequisites Part A Probability is a prerequisite. BS3a/OBS3a Applied Probability or B10 Martin- gales and Financial Mathematics would be useful, but are by no means essential; some material from these courses will be reviewed without proof. Aims L´evy processes form a central class of stochastic processes, contain both Brownian motion and the Poisson process, and are prototypes of Markov processes and semimartingales. Like Brownian motion, they are used in a multitude of applications ranging from biology and physics to insurance and finance. Like the Poisson process, they allow to model abrupt moves by jumps, which is an important feature for many applications. In the last ten years L´evy processes have seen a hugely increased attention as is reflected on the academic side by a number of excellent graduate texts and on the industrial side realising that they provide versatile stochastic models of financial markets. This continues to stimulate further research in both theoretical and applied directions. This course will give a solid introduction to some of the theory of L´evy processes as needed for financial and other applications. Synopsis Review of (compound) Poisson processes, Brownian motion (informal), Markov property. Connection with random walks, [Donsker’s theorem], Poisson limit theorem. Spatial Poisson processes, construction of L´evy processes. Special cases of increasing L´evy processes (subordinators) and processes with only positive jumps. -
Geometric Brownian Motion with Affine Drift and Its Time-Integral
Geometric Brownian motion with affine drift and its time-integral ⋆ a b, c Runhuan Feng , Pingping Jiang ∗, Hans Volkmer aDepartment of Mathematics, University of Illinois at Urbana-Champaign, Illinois, USA bSchool of Management and Economics,The Chinese University of Hong Kong, Shenzhen, Shenzhen, Guangdong, 518172, P.R. China School of Management, University of Science and Technology of China, Hefei, Anhui, 230026, P.R.China cDepartment of Mathematical Sciences, University of Wisconsin–Milwaukee, Wisconsin, USA Abstract The joint distribution of a geometric Brownian motion and its time-integral was derived in a seminal paper by Yor (1992) using Lamperti’s transformation, leading to explicit solutions in terms of modified Bessel functions. In this paper, we revisit this classic result using the simple Laplace transform approach in connection to the Heun differential equation. We extend the methodology to the geometric Brownian motion with affine drift and show that the joint distribution of this process and its time-integral can be determined by a doubly-confluent Heun equation. Furthermore, the joint Laplace transform of the process and its time-integral is derived from the asymptotics of the solutions. In addition, we provide an application by using the results for the asymptotics of the double-confluent Heun equation in pricing Asian options. Numerical results show the accuracy and efficiency of this new method. Keywords: Doubly-confluent Heun equation, geometric Brownian motion with affine drift, Lamperti’s transformation, asymptotics, boundary value problem. ⋆R. Feng is supported by an endowment from the State Farm Companies Foundation. P. Jiang arXiv:2012.09661v1 [q-fin.MF] 17 Dec 2020 is supported by the Postdoctoral Science Foundation of China (No.2020M671853) and the National Natural Science Foundation of China (No.