<<

NATIONAL TECHNICAL UNIVERSITY OF ATHENS SCHOOL OF NAVAL ARCHITECTURE & MARINE ENGINEERING Section of Naval & Marine Hydrodynamics

Stochastic Analysis with Applications to Dynamical Systems

by

Themistoklis P. Sapsis

Diploma Thesis

Supervisor: G. A. Athanassoulis, Professor NTUA

Athens September 2005 ii

iii

Preface

This thesis deals with the analysis of multidimensional nonlinear dynamical systems subject to general external stochastic excitation. The method used to address this problem is based on the Characteristic . As it is known, the characteristic functional includes the full structure of a . Hence the knowledge of the characteristic functional induces the knowledge of every statistical quantity related with the stochastic process. For Stochastic Dynamical Systems, the joint Characteristic Functional of the response and the excitation is governed by linear Functional Differential Equations, i.e. differential equations with Volterra derivatives. This kind of equations appeared for the first time at the paper “Statistical Hydromechanics and ” by E. Hopf (1952) in the context of the analysis of Turbulence. Even thought such an equation contains the full probabilistic structure of the system response cannot be a straight forward method of solution of the problem. Apparently no practical methods have been found for the efficiently solution of this type of equations. The present work deals with two very important issues, concerning Functional Differential Equations that describe Stochastic Dynamical Systems. First, we show that apparently all known equations partially describing the probability structure of the response under specific assumptions (such as Fokker-Planck Equation, Liouville Equation, Moment Equations, etc) can be derived directly through the Functional Differential Equation. Additionally, we derive some new (to the best our knowledge) Partial Differential Equations governing the characteristic functions of the response for systems under general stochastic excitation. In a second level, a new method is proposed for the solution of Functional Differential Equations. This method is related with the concepts of Kernel Probability Measures and Kernel Characteristic Functionals in infinite dimensional spaces, which are defined and discussed extensively. In this way we propose a method for treating the Functional Differential Equations at the infinite dimensional level, without any reduction to PDE’s. A simplification of the proposed method is implemented and a numerical example is presented. This thesis is divided into five chapters. In Chapter 1 we briefly present the basic background needed for our study. We recall some basic elements of set theory as well as the notions of probability and probability space. In the last section of the chapter we review the concept of random variable and give a summary of some results concerning their characteristic properties. The first part of Chapter 2 presents the concept of stochastic process as well as an extensive discussion concerning the Kolmogorov Theorem. This discussion will be very useful for Chapter 3 where we will define the stochastic process using the measure theoretic approach. The second part of chapter 2 deals with some classifications of the stochastic processes. Stochastic processes are categorized with respect to their memory, their regularity and their ergodic properties. Chapter 3 deals with the stochastic processes through the measure theoretic approach. In the first two sections we extensively discuss the properties of a probability measure defined on a as well as its finite dimensional produced measures. In Sections 3.3-3.5 the characteristic functional is defined and studied. Especially we study the connection of characteristic functional with more conventional statistical quantities describing the stochastic processes, such as moments, iv characteristic functions etc. Finally, in the three last sections of the chapter, we study specific cases of probability measures/characteristic functionals, such as the Gaussian and the Poisson. Special attention has been given to the characteristic functional introduced by V. Tatarskii, which induces as special (or asymptotic) cases almost every known characteristic functional. Before we proceed to the general study of Stochastic Dynamical Systems, we present the topic of mean-square calculus. Although classical, it’s very important for a complete description of Stochastic Differential Equations. Among the topics we examine, of special interest is the presentation for the of stochastic processes where we distinguish the case of integration of a stochastic process with respect to another independent stochastic process and the case of integration over Martingales with non-anticipative dependence. At the end of the chapter we present some results and criteria for the sample path properties of stochastic process. Finally in the beginning of Chapter 5 the notion of a stochastic differential equation is defined and some general theorems concerning the existence and uniqness of its solutions are proved. Moreover, necessary conditions for bounded of all moments of the probability measure describing the system response are given. Section 5.2 deals with the formulation of functional differential equations for the joint characteristic functional describing (jointly) the probability structure of the excitation and the response. The formulation concerns stochastic systems described by nonlinear ODE’s with polynomial nonlinearities. We distinguish between the case of a m.s.- continuous stochastic excitation and an orthogonal-increment excitation. Before the general results we present two specific systems, the Duffing oscillator and the van der Poll oscillator. In the next section 5.3 we prove how the general functional differential equation describing the probability structure of a system can produce a family of partial differential equations for the characteristic functions of various orders. Special cases are the Fokker-Planck equation, as well as the Liouville equation. Of great importance are the partial differential equations proved at Section 5.3.4 that describe the probability structure of the system for the case of m.s.-continuous excitation. In the last Section 5.5 we introduce the notion of the kernel characteristic functional that generalizes the corresponding idea of kernel density functions in infinite dimensional spaces. We prove that the Gaussian characteristic functional has the kernel property, thus being an efficient tool for the study of FDEs. The proof is based on existent theorems for extreme values of stochastic processes. In Section 5.4.4 we discuss how the kernel characteristic functionals can be efficiently used for the solution of functional differential equations. The basic idea is the localization in the physical phase space domain (probability measure) expressed in terms of the characteristic functional. It should be noted that the Gaussian measures are of great importance in this direction since most (if not all) of the existing analytical results concerning infinite dimensional pertain Gaussian measures. In this way we derive a set of partial differential equations that governs the functional parameters of the kernel characteristic functionals. Finally, in Section 5.4.5, we a simplified set of equations is derived in the basis of specific assumptions concerning the probability interchange between probability kernels. The latter equations are applied for the numerical study of a simple nonlinear ODE.

v

Acknowledgments

Acknowledgments are owed to everyone who helped and assisted me, each one on his/her own way, during the elaboration of this thesis. However, let me take this opportunity to thank some people individually. First, I would like to express my deep gratitude to the supervisor of this work, Prof. Gerassimos A. Athanassoulis, for his guidance, encouragement and support throughout the whole duration of my studies at the school of Naval Architecture and Marine Engineering of NTUA. I would like to thank him for bringing me in touch, through his inspiring teaching, with various mathematical problems and especially the examined one. This thesis could never have been written or even conceived without all the stimulating discussions that we had the last four years. The valuable contribution of Mr. Panagiotis Gavriliadis, PhD candidate, through the critical discussions as well as of Mr. Agissilaos Athanassoulis, PhD candidate, through his stimulating comments that led to a number of improvements, are greatly appreciated. Finally I have to thank my family Panagiotis, Euaggelia and Pantelis Sapsis as well as Kiriaki Kofiani, student of NTUA. This thesis could never have been taken the current form without their continuous and unconditional love and support. This thesis is dedicated to them with immense gratitude.

Th. Sapsis September 2005

vi

vii

Contents

Preface iii

Acknowledgments v

Contents vii

List of Symbols xi

Chapter 1: BACKGROUND CONCEPTS FROM PROBABILITY AND ANALYSIS

1.1. Elements of set theory 2 1.1.1. Set operations 2 1.1.2. Borel Fields 4

1.2. Axioms of Probability – Definition of Probability Measure 5

1.3. Random Variables 7 1.3.1. Distribution Functions and Density Functions 7 1.3.2. Moments 9 1.3.3. Characteristic Functions 10 1.3.4. Gaussian Random Variables and the Central Limit Theory 13 1.3.5. Convergence of a Sequence of Random Variables 16 1.3.5.a. Mean-Square Convergence 16 1.3.5.b. Convergence in Probability 17 1.3.5.c. Almost Sure Convergence 17 1.3.5.d. Convergence in Law or Distribution 17 1.3.5.e. Relationships Between the Four Modes of Convergence 17

1.4. References 18

viii

Chapter 2: STOCHASTIC PROCESSES. DEFINITION AND CLASSIFICATION

2.1. Basic Concepts of Stochastic Processes 20 2.1.1. Definition of a Continuous-parameter Stochastic Process 20 2.1.2. Moments 23 2.1.3. Characteristic Functions 24

2.2. Classification of Stochastic Process 26 2.2.1 Classification of Stochastic Process based upon Memory 26 2.2.1.a Purely Stochastic Process - Generalized Stochastic Process 26 2.2.1.b Markov Process 29 2.2.1.c Independent Increment Process 31 2.2.1.d. Martingales 38 2.2.2. Classification of Stochastic Process based upon Regularity 41 2.2.2.a Stationary and Wide-Sense Stationary Processes 41 2.2.2.b Spectral Densities and Correlation Functions 42 2.2.3. Classification of Stochastic Process based upon Ergodicity 45 2.2.4 Gaussian Process 47

2.3. References 49

Chapter 3: PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONAL IN HILBERT (FUNCTION) SPACES

3.1. Probability Measures in Hilbert Spaces and their finite-dimensional relatives 52 3.1.1. Stochastic Processes as Random variables defined on Polish Spaces 52 3.1.2. Cylindric Sets 53 3.1.3. Restriction of a probability measure over a Polish space, to finite-dimensional subspaces 54 3.1.4. Conditions for the existence of probability measures over a Polish space 55 3.2. Cylinder functions and their Integral 59 3.3. Characteristic Functional of the Probability Measure 61 3.4. Mean and Covariance Operator for Probability Measures 69 3.5. Moments from the Characteristic Functional 72 3.6. Characteristic Functions from the Characteristic Functional 75 3.7. Gaussian Measure 77 ix

3.8. Tatarskii Characteristic Functional 80 3.9. Characteristic Functionals derived from Tatarskii Functional 85 3.10. References 90

Chapter 4: – PRINCIPLES AND RESULTS

4.1. Mean-Square Calculus of 2nd -Order Process (smooth and non-smooth) 94 4.1.1. Preliminaries 94 4.1.2 Mean-Square Convergence 94 4.1.3 Mean-Square Continuity 96 4.1.4 Mean-Square Differentiation 97 4.1.5 Mean-Square Integration I – Integration of a Stochastic Process with respect to another (Independent) Process 100 4.1.6 Mean-Square Integration II – Integration over Martingale Stochastic Process with non- anticipative dependence 103 4.1.6.a A formula for changing variables (Ito Formula) 106 4.1.6.b Stochastic Differentials 107

4.2. Analytical Properties of Sample Functions 110 4.2.1. Sample Function Integration 110 4.2.2. Sample Function Continuity 111 4.2.3 Sample Function Differentiation 113

4.3. References 116

Chapter 5: PROBABILISTIC ANALYSIS OF THE RESPONCES OF DYNAMICAL SYSTEMS UNDER STOCHASTIC EXCITATION

5.1. Stochastic Differential Equations – General Theory 119 5.1.1. General Problems of the Theory of Stochastic Differential Equations 119 5.1.2. The Existence and Uniqueness Theorems of Stochastic Differential Equations 124 5.1.3. Bounds on the Moments of Solutions of Stochastic Differential Equations 128 5.1.4. Continuous dependence on a parameter of solutions of stochastic equations 129

x

5.2. Hopf-type Functional Differential Equations (FDE) for the Characteristic Functional of Dynamical Systems 133 5.2.1. Duffing oscillator 133 5.2.2. Van Der Poll Oscillator 135 5.2.3. Non-Autonomous Dynamical Systems under Stochastic Excitation I – Case of m.s. Continuous Excitation 136 5.2.4. Non-Autonomous Dynamical Systems under Stochastic Excitation II – Case of Independent Increment Excitation 140

5.3. Reduction of the Hopf-type FDE to PDEs for (joint) characteristic functions or (joint) probability densities 144 5.3.1. Systems with Random Initial Conditions 144 5.3.2. Moment Equations from the FDE 147 5.3.3. Excitation with Independent Increments - Fokker-Planck Equation 150 5.3.4. Partial Differential Equation for (N+M)-dimensional joint (excitation-response) characteristic functions in the case of general m.s. Continuous Excitation 160

5.4. Kernel representations for the Characteristic Functional 169 5.4.1 Exact Formula for Integrals of Special Functionals 169 5.4.2 Integrals of Variations and of Derivatives of Functionals 177 5.4.3 Kernel Characteristic Functionals 179 5.4.4. Superposition of Kernel Characteristic Functionals for the numerical solution of FDEs 186 5.4.5. Simplification of the set of FDE’s governing the Kernel Characteristic Functionals 194

5.5. References 201

Subjects for Future Research 203

Appendix I : Volterra Derivatives of Gaussian Characteristic Functional 205

Appendix II : Optimization 207

Appendix III : Application on Ship Stability and Capsizing on Random Sea 209

Index xi

List of Symbols

Ω Space of all possible events of a random experiment U ()Ω σ – of subsets of Ω c Probability measure on the U (Ω)

cX Induced probability measure of the random variable/stochastic process X

FX Probability distribution function of the random variable X ()β

f X Probability density function of the random variable X ()β E β ()i Mean value operator

φX Characteristic function of the random variable X ()β

N []m,σ Gaussian random variable with mean m and variance σ

I ⊆ Time interval

ΠΗ ()i Linear projection operator to the subspace H .

DI() Space of real valued functions on I ⊆ m Space of random variables/stochastic processes with finite second moment

∞ CIc ()⊂ DI () Space of infinite differentiable functions with compact support

U t()Ω current of σ − algebras

Sρ Sphere with radius ρ

Y X Characteristic functional of the probability measure c X

ii, Dual product

S Space of nuclear operators

mc Mean value element of the probability measure c

Cc Correlation operator of the probability measure c V 2 Space of twice differentiable functionals δY ()[]x z Frechet derivative of the functional Y at the direction z δY Volterra derivative of the functional Y ()i at the point t δx()t

xii

Chapter 1

Background Concepts from Probability

The mathematical theory of probability and the basic concepts of random variables form the theoretical background of the present work. The reader is assumed to be familiar with these concepts, so only some basic definitions and results are summarized in this chapter mainly for establishing a consistent system of notation, and for later reference. In Section 1.1 we recall some basic elements of set theory and give the definition of a Borel field that will be used extensively in what follows. Section 1.2 deals with the notion of the probability measure. We present the three axioms of probability and discuss the completeness of the mathematical description of a random experiment. Finally in the last section we give the definition of a general random variable and is given the various probability density functions and characteristic functions. All analytical means used for describing its probability structure are defined and briefly discussed. For more detailed discussions the reader is referred to the literature ATHANASSOULIS, G.A., (Stochastic Modeling and Forecasting of Ship Systems) and FRISTEDT, B. & GRAY. L. (A Modern Approach to ).

1 2 CHAPTER 1 BACKGROUND CONCEPTS FROM PROBABILITY

1.1. Elements of Set Theory

Events and combinations of events occupy a central place in probability theory. The mathematics of events is closely tied to the theory of sets, and this section constitutes a summary of some elements of this theory. A set is a collection of arbitrary objects. These objects are called elements of the set and they can be of any kind with any specified properties. We may consider, for example, a set of numbers, a set of points or a set of functions. A set containing no elements is called an empty set and .We distinguish between sets containing a finite number of elements and those having an infinite number. They are called, respectively, finite sets and infinite sets. An infinite set is called enumerable or countable if all its elements can be arranged in such a way that there is a one-to-one correspondence between them and all positive integers; thus, a set containing all positive integers 1,2,... is a simple example of an enumerable set. A nonenumerable or noncountable set is one where the above mentined one-to- one correspondence cannot be established. A simple example of nonenumerable set is the set of all points on a straight line segment. If every element of a set A is also an element of a set B , the set A is called a subset of B and this is represented symbolically by A ⊂ B (1) It is clear that an empty set is a subset of any set. In the case where both AB⊂ and B ⊂ A hold, the set A is then equal to B, and we write A = B (2) We now give meaning to a particular set we shall called space. In our development, we shall consider only sets which are subsets of a fixed (nonempty) set. This "largest" set containing all elements of all the sets under consideration is called the space, and it is denoted by the symbol Ω . The class of all the sets in Ω is called the space of sets in Ω . It will be denoted by _ (Ω). Consider a subset A in Ω . The set of all elements in Ω which are not elements of A is called the complement of A, and we denote it by A′. The following relations clearly hold: Ω′=∅ , ∅′= Ω , ()Α′ ′ =Α

1.1.1. Set Operations

Let us now consider operations of sets A,B,C, … which are subsets of space Ω . We are primarily concerned with addition, subtraction, and multiplication of these sets. The union or sum of A and B, denoted by A∪Β , is the set of all elements belonging to A or B or both.

The intersection or product of A and B, written as A∩Β is the set of all elements which are common to A and B.

1.1. ELEMENTS FROM SET THEORY 3

If A∩Β=∅, the sets A and B contain no common elements, and we call A and B mutually exclusive or disjoint. The symbol " + " shall be reserved to denote the union of two disjoint sets.

The definitions of the union and the intersection can be directly generalized to those involving an arbitrary number (finite or countably infinite) of sets. Thus, the set

n AA12∪∪∪… Anj= ∪ A j=1

stands for the set of all elements belonging to one or more of the sets Aj ,1,2,,jn= … . The intersection

n AA12∩∩∩… Anj= ∪ A j=1

is the set of all elements common to all Ajj ,1,2,,= … n. The sets Ajj ,= 1, 2,… , n are called mutually exclusive if

Aij∩ Aijnfor every ,= 1, 2,… , (3)

It is easy to verify that the union and the intersection operations of sets are associative, commutative, and distributive, that is

()AB∪∪ CA== ∪∪ () BC ABC ∪∪ (4)

ABBA∪∪= (5)

()AB∩∩ CA== ∩∩ () BC ABC ∩∩ (6)

ABBA∩∩= (7)

ABCAB∩∪()= ( ∩∪∩)( AC ) (8)

Clearly, we also have

A∪ AA= , A∪∅=A , A∩∅ =∅

A∪ΩΩ= , AA∩Ω= , AA∩ ′ = ∅

Moreover, the following useful relations hold

A∪∩()BC= ( AB ∪∩∪)( AC ) (9)

()AB∩∪′ = A′′ B (10)

A∪∪∩BAA= ′ B (11)

The last two relations are referred to as De Morgan's law.

4 CHAPTER 1 BACKGROUND CONCEPTS FROM PROBABILITY

Finally, we define the difference between two sets. The difference AB− is the set of all elements belonging to A but not to B . From the definition, we have the simple relations A −∅= A , Ω −=A A′ , AB−= A∩ B′

1.1.2. Borel Fields

Definition 1.1 : Given a space Ω , a Boret field, or a σ – field U of subsets of Ω is a class of, in general noncountable, subsets Ajj ,= 1, 2,… , having the following properties: 1. Ω∈U 2. If A∈U then A′∈U ∞ 3. If Aj ∈=U ,1,2,,j … then ∪ Aj ∈U . ■ j=1 The first two properties imply that ∅ −∈Ω′ U (12) and, with the aid of De Morgan's law, the second and the third properties lead to ∞∞⎛⎞′ ∩∪AAjj= ⎜⎟′ ∈U (13) jj==11⎝⎠ A Borel field is thus a class of sets, including the empty set ∅ and the space Ω , which is closed under all countable unions and intersections of its sets. It is clear, of course, that a class of all subsets of Ω is a Borel field. However, in the development of the basic concepts of probability, this particular Borel field is too large and impractical. We in general consider the smallest class of subsets of Ω which is a Borel field and it contains all sets and elements under consideration.

1.2. AXIOMS OF PROBABILITY – DEFINITION OF A PROBABILITY MEASURE 5

1.2. Axioms of Probability – Definition of Probability Measure

The basic concepts of probability theory are centered around the idea of a random experiment E , whose outcomes are events. Given a random experiment E the collection of all possible events is called a sample space, whose elements are the simple events. Observable events enter the sample space as its subsets. The definitions of events and sample space provide a framework within which the analysis of events can be performed, and all definitions and relations between events in probability theory can be described by sets and set operations in the theory of sets. Consider the space Ω of elements ω and with subsets AB , ,…

In our subsequent discussion of probability theory, we shall assume that both Ω and ∅ are observable events. We also require that the collection of all observable events associated with a random experiment E constitutes a Borel field U which implies that all events formed through countable unions and intersections of observable event are observable, and they are contained in U .

We now introduce the notion of a probability function. Given a random experiment E , a finite number c ()A is assigned to every event A in the Borel field U of all observable events. The number c ()A is a set function and is assumed to be defined for all sets in U . We thus have the following

Definition 2.1 : Let the space Ω of elements ω and the Boret field U ()Ω generated by the open subsets of Ω . The set function cU:0,1(Ω) → [ ] will be a probability function or probability measure iff the following axioms hold

1. c ()A ≥ 0 for every A∈U (Ω) 2. c ()Ω =1 3. for any countable collection of mutually disjoint sets ⎪⎪⎧⎫ AA12, ,..., inUc()Ω : ⎨⎬∪ Ajj= ∑ c() A ■ ⎩⎭⎪⎪j j

Hence, Axioms 1-3 define a countably additive and nonnegative set function cU()AA, ∈ ()Ω . The following properties associated with the probability function can be easily deduced from the axioms stated above

c ()∅=0 (1)

If A ⊂ B , then

cc()AB≤ () (2)

We also see that in general,

cccc()AB∪∩=+−() A () B( AB) (3)

6 CHAPTER 1 BACKGROUND CONCEPTS FROM PROBABILITY

A mathematical description of a random experiment is now complete. As we have seen above, it consists of three fundamental constituents: A sample space Ω , a Borel field U (Ω) of observable events, and the probability function cU:0,1(Ω) → [ ] . These three quantities constitute a probability space associated with a random experiment, and it is denoted by ()Ω,,Uc .

1.3. RANDOM VARIABLES 7

1.3. Random Variables

Consider a random experiment E whose outcomes β are elements of Ω in the underlying probability space ()Ω,,Uc . These outcomes are in general represented by real numbers, functions, or objects of various kinds. In order to construct a model for a random variable, we assume that for all experiments it is possible to assign a real number , function, or other object of various kinds X ()ω for each ω following a certain set of rules. The "number" X ()ω is really a point function defined over the domain of the basic probability space. We have the following

Definition 3.1 : Let (Ω,,Uc) be a probability space, X :Ω→X a point function and

UX a Borel field over X . The mapping X is a random variable, iff is ()UU, X - measurable. ■ In the above definition the probability space (Ω,,Uc) is modelling a stochastic experiment and the set X is modelling the quantities, at those we focus our attention. The mapping X :Ω→X is modelling the procedure of taking the quantity X ()ω . The above value is a realization of the random variable.

Definition 3.2 : Let ()Ω,,Uc be a probability space, ()X , UX a measurable space and

X :Ω→X a random variable. On the Borel field UX of the measurable space ()X , UX we define the induced probability measure

−1 ccXX()BXB=∈() () ,for every B U (1)

The produced probability space ()X ,,UcXX, will be called as X −induced probability space. ■

1.3.1. Distribution Functions and Density Functions

Given a random experiment with its associated r.v. X ()β and given a real number x, let us consider the probability of the event {}ββ: X ()≤ x , or simply c X ()X ()β ≤ x . This probability clearly depends, upon x . The function

FXX()x =≤c ()Xx (β ) (2) is defined as the distribution function of X ()β . In the above, the subscript X identifies the r.v. The distribution function always exists. By definition, it is a nonnegative, continuous to the right, and nondecreasing function of the real variable x . Moreover, we have

FX ()−∞ = 0 and FX ()∞=1 (3)

8 CHAPTER 1 BACKGROUND CONCEPTS FROM PROBABILITY

A r.v. X ()β is called a continuous r.v. if its associated distribution function is continuous and differentiable almost everywhere. It is a discrete r.v. when the distribution function assumes the form of a staircase with a finite or countably infinite jumps. For a continuous r.v. X ()β , the derivative

dF ()x fx()= X (4) X dx in this case exists and is called the density function of the r.v. X. It has the properties,

fxX ()≥ 0 (5)

b f xdx=− b a (6) ∫ XXX()FF () () a

∞ fxdx=1 (7) ∫ X () −∞

On the other hand, the density function of a discrete r.v. or of a r.v. of the mixed type does not exist in the ordinary sense. However, it can be constructed with the aid of the . Consider the case where a r.v. X takes on only discrete values x12,,,xx… n . A consistent definition of its density function is

n f Xjj()xpxx=−∑ δ() (8) j=1 where

p j ==c Xj()Xx()β (9)

This definition of f X ()x is consistent in the sense that it has all the properties indicated by equations (5)-(7), and the distribution function FX ( x) is recoverable from equation (4) by integration.

The definition of the distribution function and the density function can be readily extended to the case of many random variables. Thus, the joint distribution function of a sequence of n r.v.'s, {}X n ()β , is defined by

F x ,,,xx……=≤≤≤ XxXx Xx (10) XX12…… Xnn()12 nc XX 12 X( 1 1∩∩∩ 2 2 n n)

The corresponding joint density function is

fFx ,,,xx………=∂n xxxxxx ,,, / ∂ ∂ ∂ (11) XX12…… Xnn()12 n XX 12 X( 12 n) 12 n

if the indicated partial derivative exists. The finite sequence {}X n ()β may be regarded as the components of an n -dimensional random vector X ()β . The distribution function and the density function of X ()β are identical with (10) and (11), but they may be written in the more 1.3. RANDOM VARIABLES 9

compact forms FxX () and f X ()x , where x is a vector with components x12,,,xx… n . In the sequel, the existence of the density functions shall always be assumed.

1.3.2. Moments

Some of the most useful information concerning a random variable is revealed by its moments, particularly those of the first and second order. The nth moment of a r.v. X ()β is defined by

⎧ n ⎪∑ xpii for discrete r.v. n ⎪ i EXβ β = (12) (()()) ⎨+∞ ⎪ xfn xdxfor continuous r.v ⎪ ∫ X () ⎩−∞ +∞ if xn fxdx is finite. The first moment EXβ β gives the statistical average of the r.v. ∫ X () ( ( )) −∞

X ()β . We will usually denote it by mX .

The central moments of a r.v. X ()β are the moments of X ()β with respect to its mean. Hence, the nth central moment of X ()β , is defined as

⎧ n ⎪∑()xmiXi− p for discrete r.v. n ⎪ i EXβ β −= m (13) (()() X ) ⎨+∞ ⎪ xm− n f xdxfor continuous r.v ⎪ ∫ ()()XX ⎩−∞

The moments of two or more random variables are defined in a similar fashion. The joint moments of the r.v.'s X ()β and Y ()β , are defined by

⎧ nn ⎪∑ xyijij p for discrete r.v. nm⎪ ij, EXβ ββ⋅= Y (14) (()() ()() ) ⎨+∞ ⎪ xnm⋅⋅ y f x dx dy for continuous r.v ⎪ ∫ XY () ⎩−∞

The central joint moments of the r.v.'s X ()β and Y ()β , are defined by

⎧ n m ⎪∑()xmiX−−() y jYij m p for disc. r.v. nm⎪ ij, EXβ ββ−−= m Y m (15) (()() XY()() ) ⎨+∞ ⎪ x−− mnm y m f x dx dy for cont. r.v ⎪ ∫ ()()()XYXY ⎩−∞ when they exist.

For nm==1 we have the covariance function β REXYXY =⋅()()β ()β (16)

10 CHAPTER 1 BACKGROUND CONCEPTS FROM PROBABILITY

and the correlation function β CEXXY=−⋅−()()()ββ mY X()() m Y (17) Obviously, we have

CRmmXY=− XY X Y (18)

The r.v.'s X ()β and Y ()β , are said to be uncorrelated if

CXY = 0 (19) which implies

EXβββ()()β ⋅= Y ()βββ EX()() ⋅ EY( ( )) (20)

In the particular case when

EXβ ()()ββ⋅= Y () 0 (21) the r.v.'s are called orthogonal.

It is clearly seen from Eq. (18) that two independent r.v.'s with finite second moments are uncorrelated. We point out that the converse of this statement is not necessarily true. In closing this subsection, we state two inequalities involving the moments of random variables which will be used extensively in the latter chapters.

Tchebycheff Inequality. n EXβ ()()β c ()Xk()β ≥≤ (22) k n Hölder Inequality. 11 nmnm EXβββ()ββ⋅≤ Y () ⎡ EX() β⎤⎡ ⋅ EY() β ⎤ (23) ()⎣⎢ ( )⎦⎣⎥⎢( ) ⎦⎥ where, 11 nm,1> , with +=1 nm The proof of the above inequalities can be found at KOLMOGOROV, A.N. & FOMIN, S.V, (Introductory Real Analysis).

1.3.3. Characteristic Functions

Definition 3.3 : Let X ()β be a r.v. with density function f X ()x . The characteristic function of X ()β , φX ()u is defined by

⎪⎧ ⎪ epiux j for discrete r.v. ⎪∑ j ⎪ j φ ()uEe==β iuX ⎨⎪ (24) X ()⎪ ∞ ⎪ efiux () xdxfor continuous r.v ⎪∫ X ⎩⎪−∞ 1.3. RANDOM VARIABLES 11

It is seen that the characteristic function φX ()u and the density function f X ()x form a Fourier transform pair. Since a density function is absolute integrable, its associated characteristic function always exists. Furthermore, it follows from the theory of Fourier transforms that the density function is uniquely determined in terms of the characteristic function by 1 ∞ f ()xeudu= −iuxφ () (25) XXπ ∫ 2 −∞ Eq. (25) points out one of the many uses of a characteristic function. In many physical problems, it is often more convenient to determine the density function of a random variable by first determining its characteristic function and then performing the Fourier transform as indicated by equation (25). Another important property of the characteristic function is its simple relation with the moments. The MacLaurin series of φX ()u gives u2 φφφ()uuuu=+ ()0 ′′′ () + φ () +… (26) XXX X2 Now, we see that ∞ φ 01==fxdx XX()∫ () −∞ ∞ φβ′ 0 ==ixfxdxiEXβ XX()∫ ()() ( ) −∞ ∞ n φβ()n 0 ==ixfxdxiEXnn nβ (27) XX()∫ ()()() ( ) −∞

The joint characteristic function of many r.v.’s {}X i ()β is defined by

⎛⎞n ⎜ iuX∑ jj⎟ β ⎜ j=1 ⎟ φ ()uu,,,… u== E⎜ e ⎟ XX12… Xn 12 n ⎜ ⎟ ⎜ ⎟ ⎝⎠ (28) n ∞∞ iux∑ jj = ………efj=1 () xxxdxdx,,, ∫∫ XX1… n 12 n 1 n −∞ −∞

Analogous to the one-random-variable case, the joint density function f xx,,,… x is XX12… Xn ( 12 n) uniquely determined in terms of φ ()uu,,,… u by the n-dimensional Fourier transform XX12… Xn 12 n fxxx(),,,… = XX12… Xn 12 n n ∞∞ ∞ 1 iux∑ jj (29) = ………ej=1 φ () u,,, u u du du du n ∫∫ ∫ XX12… Xn 12 n 1 2 n ()2π −∞ −∞ −∞ The moments, if they exist, are related to φ ()uu,,,… u as XX12… Xn 12 n

12 CHAPTER 1 BACKGROUND CONCEPTS FROM PROBABILITY

mm m β ββ12… βn = EX()()12() () X() () Xn () ++… (30) 1 ∂mm1 n =⋅++… φXX… X ()0,0,… ,0 iuumm1 nn∂∂m1 … m12 n Since a density function is absolute integrable, its associated characteristic function always exists. Here are some of the basic properties of the characteristic function:

1) φ()0,0,… ,0= 1 2) φφ()uu=− ( ) 3) φ()u is uniformly continuous for each u ∈ n 4) φ()u is a positive–definite function

Concerning the characterization of characteristic functions, there is the following

Theorem 3.4 [Bochner] : The function g : n → satisfies the above conditions if and only if there exists an n-dimensional random vector X ()β , whose characteristic function is g()u .

Proof : The “if” assertion is immediate, since, if g is a characteristic function with density function G , then, letting u and υ range over an arbitrary but finite set in R ,

∑∑gu()()()−=υυ huh∫ eiu( −υ) x huh()() υ dGx() = uu,,υυ 2 =≥∫ ∑ ehudGxiux () () 0 u

Conversely, let g on R be nonnegative-definite and continuous. It coincides on R with a characteristic function. If it does do on the set Sr (dense in R ) of all rationals of the form of n the form k /2 , k =±±0, 1, 2,… , n = 1,2,… . For every integer n , let Sn be the n corresponding subset of all rationals of the form k / 2 so that SSnr↑ . Since g is nonnegative- definite on R , it is nonnegative-definite on every Sn . Therefore, by Herglotz Lemma (see nn LOÈVE, Probability Theory) there exists characteristic functions such that gk()()/2= fn k /2 whatever be k and n . Since SSnr↑ , it follows that fn → g on Sr . Let 0 ≤ θ , θn ≤1, so that, by Herglotz Lemma,

π 1Re−=−≤fxdFxθθ /2nn 1cos 2 nn()∫ () () −π π ≤−1cosxdF 2nn x =− 1Re1/2 f ∫ n () ( ) −π

Therefore, by the elementary inequality ab+≤22222 a + b and the increments inequality, n for fixed hk=+()nnθ /2 ,

1.3. RANDOM VARIABLES 13

2 2 nn 121/241Re/2−≤−fhnnnnn() fk() +−() f()θ ≤

nn ≤−21gk()n /2 + 4() 1 − Re g() 1/2

Since g is continuous at the origin, it follows that the sequence fn is equicontinuous. Hence, by Ascoli’s Theorem, it contains a subsequence converging to continuous function f so that

gf= on Sr and hence on R . Since by the continuity theorem f is a characteristic function, the proof is complete.

The “only if” assertion can be proved directly, and this direct proof will extend to a more general case: For every T > 0 and x ∈ R

1 TT p x=− g uυυ e−−iu()υ x dud ≥0 T () ∫∫ () T 00 since, g on R being nonnegative-definite and continuous, the integral can be written as a limit of nonnegative Riemann sums. Let ut=+υ , integrate forst with respect to υ and set ⎛⎞t ⎜ ⎟ gtT ()=−⎜1 ⎟ gt() or 0 according as tT≤ or tT≥ . The above relation becomes ⎝⎠⎜ T ⎟

1 px()=≥ gtedt() −itx 0 TTT ∫

⎛⎞ 1 ⎜ x ⎟ iux Now multiply both sides by ⎜1− ⎟e and integrate with respect to x on ()−X , X . The 2π ⎝⎠⎜ X ⎟ relation becomes

⎛⎞ 2 ⎜1 ⎟ X sin ⎜ Xt()− u⎟ 11⎛⎞x ⎝⎠⎜2 ⎟ ⎜1−=⎟ p xeiux g tdt ∫∫⎜ ⎟ TT() () 22ππ⎝⎠⎜ X ⎟ 1 2 −X Xt()− u 4

The l.h.s. is a characteristic function (since its integrand is a product of eiux by a nonnegative function) and the r.h.s. converges to guT ( ) as X →∞. Therefore, gT is the limit of a sequence of characteristic functions. Since it is continuous at the origin, the continuity theorem applies and gT is a characteristic function. Since ggT → as T →∞, the same theorem applies, and the assertion is proved. ■

An extensive discussion for the characteristic functions and their properties can be found at ATHANASSOULIS, G.A., (Stochastic Modeling and Forecasting of Ship Systems) and LOÈVE, (Probability Theory, p.198-213).

1.3.4 Gaussian Random Variables and the Central Limit Theory

The most important distribution of random variables which we encounter in physical problems is the Gaussian or .

14 CHAPTER 1 BACKGROUND CONCEPTS FROM PROBABILITY

Definition 3.5 : A r.v. X ()β is Gaussian or normal if its density function f X ()x has the form

⎧⎫2 1 ⎪⎪()xm− fx()=−exp⎨⎬⎪⎪ ■ (31) X σπ ⎪⎪2σ 2 2 ⎩⎭⎪⎪

Its distribution function is the integral

x ⎧⎫2 1 ⎪⎪()tm− Fx()=−exp⎨⎬⎪⎪ dt (32) X ∫ σπ2 ⎪⎪2σ2 −∞ ⎩⎭⎪⎪

It is seen that a Gaussian r.v. is completely characterized by the two parameters m andσ ; they are, as can be easily shown, the mean and the standard deviation of X ()β , respectively. Usually we donote it as N []m,σ .

Alternatively, a Gaussian r.v. can be defined in terms of its characteristic function. It is of the form

⎪⎪⎧⎫1 22 φσX ()uiumu=−exp⎨⎬ (33) ⎩⎭⎪⎪2

As indicated by equation (30), the moments of X ()β can be easily derived from equation (33) by successively differentiating φ(u) with respect to u and setting u = 0 . The results, expressed in terms of its central moments, are

n ⎪⎧0,n odd EXβ β −= m (34) (()() X ) ⎨ n ⎩⎪1⋅⋅ 3… ⋅()nn − 1 ⋅σ , even

In addition to having these simple properties, another reason for the popularity of a Gaussian r.v. in physical applications stems from the central limit theorem stated below. Rather than giving the theorem in general terms, it serves our purpose quite well by stating a more restricted version due to Lindeberg.

Theorem 3.6 [Central Limit Theorem] : Let {}X n ()β be a sequence of mutually independent and identically distributed r.v.'s with means m and variances σ 2 . Let

n XX()ββ= ∑ j () (35) j=1 and let the normalized r.v. Y ()β be defined as

X ()β −nm Y ()β = 1 (36) σn 2

Then the distribution function of Y ()β , FyY () , converges to the zero mean, unit variance Gaussian distribution as n →∞ for every fixed y .

Proof : We first remark that the r.v. Y ()β is simply the "normed" r.v. X ()β with mean zero and variance one. In terms of the characteristic functions φX ()u of the r.v.'s X j ()β the characteristic function φY ()u of Y ()β has the form 1.3. RANDOM VARIABLES 15

11 β iuY ⎡⎤22 φσφσYX()uEe==−()⎢⎥exp iumn un (37) ⎣⎦⎢⎥()()

In view of the expansion (26), we can write

n ⎪⎪⎧⎫⎡⎤22 2 ⎪⎪11⎢⎥()σ + m ⎛⎞iu φσσu=−⎪⎜⎪exp ium n22 1 + ium n +⎜ ⎟ +=… Y () ⎨⎬()⎢⎥⎜ 1 ⎟ ⎪⎪⎢⎥2 ⎝⎠⎜ 2 ⎟ ⎩⎭⎪⎪⎣⎦σn n ⎡⎤⎛⎞uu22 ⎛ ⎞ =−⎢⎥12uno2 +⎜⎜⎟⎟ → exp − (38) ⎢⎥⎜⎜⎟⎟ ⎣⎦⎢⎥⎝⎠n ⎝2 ⎠ as n →∞. In the last step we have used the elementary identity in calculus,

lim() 1+=cnn ec n→∞ for any real c. Equation (38) shows that φY ()u approaches the characteristic function of a zero mean and unit variance Gaussian distribution in the limit. The proof is thus complete. ■

The central limit theorem thus indicates that often a physical phenomenon has a Gaussian distribution when it represents the sum of a large number of small independent random effects. This indeed is a model or at least an idealization of many physical phenomena we encounter.

Consider a sequence {}X n ()β of n r.v.'s. They are said to be jointly Gaussian if its associated joint density function has the form ffxxx()x == (,,,… ) X XX12… Xn 12 n

−−nT/2 1/2 ⎡ 1 ⎤ =−−−()2expπΛ⎢ ()()xm Λ−1 xm⎥ (39) ⎣⎢ 2 ⎦⎥ where

mT ==mm…… m⎡⎤ Eββ Xββ E X E β X β (40) []()12nn⎣⎦⎢⎥() 1() 2() ()() and Λμ= ⎡⎤ is the nn× covariance matrix of X β with ⎣⎦⎢⎥ij ()

μββ=−−EXβ ⎡⎤ mX⎡⎤ m (41) ij()⎣⎦ i() i⎣⎦⎢⎥ j () j

Again, we see that a joint Gaussian distribution is completely characterized by the first- and second-order joint moments.

Parallel to our discussion of a single Gaussian r.v., an alternative definition of a sequence of jointly Gaussian distributed random variables is that it has the joint characteristic function

⎪⎪⎧⎫TT1 φΛX ()umuuu=−exp⎨⎬i (42) ⎩⎭⎪⎪2

This definition is sometimes preferable since it avoids the difficulties when the covariance matrix Λ becomes singular. The joint moments of X ()β can be obtained by differentiating the joint characteristic function φX ()u and setting u = 0 .

16 CHAPTER 1 BACKGROUND CONCEPTS FROM PROBABILITY

It is clear that, since the joint moments of the first- and second-order completely specify the joint Gaussian distribution, these moments also determine the joint moments of orders higher than two. We can show that, if the r.v.'s {}X n ()β have zero means, all odd order moments of these r.v.'s vanish and, for n even

EXXXββββ()……=⋅⋅⋅ EXX EXX EX X (43) 12 nmmmmmm∑ ()12()( 23nn− 1 ) mm12,,,… mn

The sum above is taken over all possible combinations of n /2 pairs of n r.v.'s. The number of terms in the summation is 135⋅⋅⋅… ⋅(nn − 3) ⋅( − 1).

1.3.5. Convergence of a Sequence of Random Variables

In this section we recall the different types of convergence of a sequence of random variables. These results will be used in the following chapters.

1.3.5.a. Mean-Square Convergence

Let {}Xnn , ∈ be a sequence of second-order random variables. The sequence {}Xnn , ∈ converges in mean-square to a second-order random variable X as n →∞ iff

1 β 2 2 limEX[]n −= X 0 (44) n→∞ ()

Preposition 3.7 : The sequence {}Xnn , ∈ , of second-order random variables converges in mean-square to a second-order random variable X such that EXβ ()2 = x, x > 0 if and only if

β lim EXX()n ⋅=′ x (45) nn, ′→∞ n

Proof : Assume that eq. (45) is true. Then,

2 ββββ⎡⎤−= ⋅+⋅−⋅⋅→ EX(⎣⎦nnnn Xnnnn′′′′) EXX() EXX()()20 EXX

as mm, ′ →∞. Therefore {}Xnn , ∈ is a Cauchy sequence. But since the space of second- order random variables is complete (See LOÈVE, Probability Theory I), any Cauchy sequence of second-order random variables convergences to a second-order random variable. Hence the β sequence {}Xnn , ∈ convergences and we have lim EXX()n ⋅=′ x. nn, ′→∞ n

Conversely, if the sequence {}Xnn , ∈ converges to a second-order random variable, then we have eq. (45) because any convergent sequence is a Cauchy sequence and therefore 2 β ⎡⎤−→′ →∞ EX(⎣⎦n Xn′ ) 0, asnn . ■

1.3. RANDOM VARIABLES 17

1.3.5.b. Convergence in Probability

A sequence {}Xnn , ∈ of random variables defined on a probability space ()Ω,,Uc , convergences in probability, to a random variable X as n →∞, iff for every ε > 0,

limc XXn ()ββε−≥= () 0 (46) n→∞ ()

Convergence in probability is also called stochastic convergence or convergence in measure.

Note that XXn −≥ε denotes the subset of U such that {}ωββε∈Ω: XXn () − () ≥ .

1.3.5.c. Almost Sure Convergence

A sequence {}Xnn , ∈ of random variables defined on a probability space ()Ω,,Uc , convergences almost surely (a.s.), to a random variable X as n →∞, iff

c limXXn ()ββ≠= () 0 (47) ()n→∞

This means that subset Ω=0 βββ ∈Ω:limXXn () ≠ () is c −negligible : c ()Ω=0 0 . {}n→∞ We can also write eq. (47) as

c limXXn ()ββ== () 1 (48) ()n→∞ Convergence c −a.s. is also called convergence almost everywhere, or almost certain convergence, or convergence with probability one.

1.3.5.d. Convergence in Law or Distribution

A sequence {}Xn, ∈ of random variables with probability distributions F , convergences n X n in law (or in probability distribution) to a random variable X with probability distribution FX if the sequence of bounded positive measures F convergences weakly to F , i.e. if for all {}X n X real continuous functions f on such that fx()→ 0 as x →∞, we have

lim f ()xFXX ( dx )= f () xF ( dx ) (49) n→∞ ∫∫n

1.3.5.e. Relationships Between the Four Modes of Convergence

Almost Sure Convergence

Convergence in Convergence in Probability Law

Mean-Square Convergence

18 CHAPTER 1 BACKGROUND CONCEPTS FROM PROBABILITY

1.4. References

ASH, R.B., 1972, Measure, Integration and . Academic Press. ASH, R.B., 2000, Probability and Measure Theory. Academic Press. ATHANASSOULIS, G.A., 2002, Stochastic Modeling and Forecasting of Ship Systems, Lecture Notes NTUA. FRISTEDT, B. & GRAY. L., 1997, A Modern Approach to Probability Theory, Birkhäuser. KLIMOV, G., 1986, Probability Theory and Mathematical Statistics, MIR Publishers. KOLMOGOROV, A.N. & FOMIN, S.V, 1975, Introductory Real Analysis. Dover Publications. LOÈVE, M., 1977, Probability Theory I, Springer. PROHOROV, YU.V. & ROZANOV, YU.A., 1969, Probability Theory. Springer. SOBCZYK, K., 1991, Stochastic Differential Equations, Kluwer Academic Publishers. SOIZE, C., 1994, The Fokker-Planck Equation and its Explicit Steady State Solutions, World Scientific. SOONG, T.T., 1973, Random Differential Equations. Academic Press. VULIKH, B.Z., 1976, A Brief Course in the Theory of Functions of a Real Variable. MIR Publishers.

Chapter 2

Stochastic Processes – Definition and Classification

Random variables or random vectors are adequate for describing results of random experiments which assume scalar or vector values in a given trial. In many physical applications, however the outcomes of a random experiment are represented by functions depending upon a parameter. This parameter can be either time or spatial variable, depending on the application. Typical examples of the first case are ODEs with inherent randomness or random excitation and these cases will mainly concern us in this thesis. In the present chapter we will present the basic definitions and theorems for the description of the above mathematical objects as well as the fundamental theorems that ensure their existence. Moreover we will make a basic classification of the stochastic processes in relation with fundamental properties such as memory and ergodicity. More specifically in Section 2.2.1 the stochastic processes are classified according to the manner in which their present state is dependent upon their past history. We thus have the purely random processes without any memory. In order to obtain a mathematic satisfactory theory of theses process we introduce the concept of generalized stochastic processes. The next step in this kind of classification is the Markov processes that can be described completed form the second order quantities. Moreover, we emphasize the case of independent increment processes and martingale stochastic processes that play important role to the theory of stochastic dynamical systems. In the next Section 2.2.2 the concept of stationarity is presented. We formulate the basic definitions as well as all the necessary results for the direction of stochastic dynamical systems. Stationary stochastic processes are very useful for the description of linear systems since, under suitable assumptions, we can formulate the whole problem in terms of the spectral densities. Section 2.2.3 deals with the important issue of ergodicity, fundamental property for dynamical systems, that relates statistical or ensemble averages of a stationary stochastic process to time averages of its individual sample functions The interchangeability of ensemble and time averages has considerable appeal in practice since we can compute statistical averages of a stochastic process in terms of a long single observation of one sample function. Finally in the last Section 2.2.4 we present the Gaussian stochastic process with its basic properties.

19 20 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

2.1. Basic Concepts of Stochastic Processes

The foregoing description of a stochastic process is based upon the axiomatic approach in probability theory, originally due to Kolmogorov. Let us call each realization of a given random experiment a sample function or a member function. This description suggests that a s.p. X ()t ;β , t∈⊆ I , is a family of sample functions of the variable t, all defined on the same underlying probability space ()Ω,,Uc . Intuitively, this sample theoretic approach to the concept of a stochastic process is physically plausible since a stochastic process describing a physical phenomenon generally represents a family of individual realizations. To each realization there corresponds a definite deterministic function of t. However, a difficult point arises. A stochastic process defined this way is specified by the probability of realizations of various sample functions. This leads to the construction of probability measures on a functional space and it generally requires advanced mathematics in measure theory and functional analysis. We will consider this approach extensively in Chapter 3. On the other hand, Definition 1.1 presents a simpler approach to the concept of a stochastic process which is sufficient for the subsequent development.

2.1.1. Definition of a Continuous-parameter Stochastic Process

At a fixed t, a s.p. X ()t ;β ,t∈⊆ I , is a random variable. Hence, another characterization of a stochastic process is to regard it as a family of random variables, say,

Xt()()12;,ββ Xt ;,…depending upon a parameter tI∈⊆ . The totality of all the random variables defines the s.p. X ()t ;β . For discrete-parameter stochastic processes, this set of random variables is finite or countably infinite. For continuous-parameter processes, the number is noncountably infinite. Thus, we see that the mathematical description of a stochastic process is considerably more complicated than that of a random variable. It, in fact, is equivalent to the mathematical description of an infinite and generally noncountable number of random variables.

We recall from Section 1.3 of Chapter 1 that a finite sequence of random variables is completely specified by its joint distribution or joint density functions. It thus follows from the remarks above that a s.p. X ()t ;β ,t∈⊆ I , is defined by a family of joint distribution or density functions, taking into account that we are now dealing with infinitely many random variables.

Let us now make this definition more precise. The index set I usually is the real line or an interval of the real line. But the above definition can be extended to the case of an index set with n-dimensions.

Definition 1.1 : Let a family of r.v.’s Xt( i ;β),tIi ∈ such that for every finite set {}tt12,,… , tn the corresponding set of r.v.'s XXt11()ββ= (;, ) XXtXXt22()ββββ== (;,, )… nn () ( ;, )has a well defined joint probability distribution function 2.1. DEFINITION OF STOCHASTIC PROECESSES 21

FxxxXtx, ,...,=< ,ββ ,..., Xtx , < (1) tt12, ,..., tn () 1 2 nc ()( 1) 1∩∩( n) n that satisfies the following conditions (Kolmogorov compatibility conditions):

a) the symmetry condition: if {i1 , i2 ,..., in } is a permutation of numbers 1, 2,..., n then for arbitrary n ≥1

Ftt, ,..., t()xx i, i ,..., x i= F tt , ,..., t() xx 1 , 2 ,..., x n (2) ii12 in 12 n 12 n

b) the consistency condition: for m < n and arbitrary ttImn+1,..., ∈

F xxx, ,..., ,∞∞= ,..., F xxx , ,..., (3) ttt11,...,mm ,+ ,..., t n()() 1 2 m ttt 12 , ,..., n 1 2 n

Then this family of joint distribution functions defines a s.p. X ()t ;β ,tI∈ . ■

In the special case where F xx, ,..., x is differentiable we can alternative use the tt12, ,..., tn () 1 2 n corresponding probability density function defined by

∂n Fxxx( , ,..., ) fxxx, ,..., = tt12, ,..., tn 1 2 n (4) tt12, ,..., tn () 1 2 n ∂⋅∂⋅⋅∂xx12... xn

This family of probability density functions satisfies the following conditions:

a) the positiveness condition

fxxx, ,...,≥ 0 (5) tt12, ,..., tn () 1 2 n

b) the normalization property

∞∞∞ ⋅⋅⋅f() x, x ,..., x ⋅ dx ⋅ dx ⋅ ... ⋅ dx = 1 (6) ∫∫∫tt12, ,..., tn 1 2 n 1 2 n −∞ −∞ −∞

Based upon the definition above, another convenient and often useful method of specifying a stochastic process is to characterize it in terms of an analytic formula as a function of containing random variables as parameters. Let AA12()ββ,,, ()… An () βbe a set of random variables. A s.p. X ()t ;β characterized this way has the general form

X ()tgtAAA;;,,,β = (12… n ) (7) where the functional form of g is given.

Example 1.2 [Pierson – Longuet Higgins Model] : Representation (7) can be applied for the description of random wind generated, water waves at a specific point. In this case we can determine experimentally the spectral distribution, i.e., the amplitude that corresponds to every frequency. Hence, the stochasticity is introduced from the phase of every harmonic component. More specifically we use the following representation for the random field at a specific point

N Xt();cosβωϕβ=+∑ Aiii() t () (8) i=1

22 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

where (Aiii,,1,,ω ) = … N are given and ϕβi (),1,,iN= … are independent random variables, uniformly distributed at []0, 2π . We must note that representation (8) is consistent with the theory of linear water waves. Moreover it can be easily proved that the above stochastic process is stationary and ergodic. Additionally for large number of harmonic components N its probability distribution functions follow the Gaussian law. ■

As we mention at the beginning a stochastic process can also specified by the probability of realizations of various sample functions. For the following definitions we assume that [ is a measurable, metric space. Definition 1.3 : Let (Ω, Uc()Ω , ) be a probability space, X :Ω→[ a point function from the space of elementary events Ω to a metric space [ and U[( ) a Borel field over [ . The mapping X is a stochastic process, iff is ( UU(Ω), ([ )) -measurable. ■

Together with the definition of the stochastic process we have, as in the case of random variables the induced probability measure.

Definition 1.4 : Let ()Ω, Uc()Ω , be a probability space, ([[, U ( )) a measurable space and X :Ω→[ a stochastic process. On the Borel field U ()[ of the measurable space ()[U[, () we define the induced probability measure

−1 cc[ ()BXB=∈( ()) ,for every B U[( ) (9)

The produced probability space ()[[,,Uc() [ , will be called as X −induced probability space. ■

To connect the above probability measure defined on the infinite dimensional space ()[U[, () with the joint probability distribution functions we have the following definition.

Definition 1.5 : Let (Ω, Uc()Ω , ) be a probability space, X :Ω→[ a stochastic process and ([[,,Uc( ) [ ) the induced probability space. Let ΠN be a projection of ⎡⎤ [ with dimΠN ⎣⎦[ = N . For every ΠN we introduce the following probability measure defined on ()NN, U (),

BBxxBB==∈∈∈ΠΠ−1 : , N ■ (10) ccPNN () [[()[ ] c[({ N[ ] }) U()

Example 1.6 : For the case of [ = CI( ) being the space of real, continuous functions of a ∈ Π ⎡ ⎤ = … real variable tI, a very important measure is for JN⎣x()txtxtxt⎦ () (12 ),,, ( ) ( ) for every

Jttt=⊂{}12,,… ,N I. For this projection operator we have

X BxCIxBB=∈ii:Π ⎡⎤ ∈∈, N (11) ccΠJ () CI()(){ () ( )J ⎣⎦ () } U( )

2.1. DEFINITION OF STOCHASTIC PROECESSES 23

N Now, lets restrict the family of B ∈U ( ) to those of the form B =−∞()(),,xx12 ×−∞ ×

…×−∞(), xN . It is easy to see that we get the joint probability distributions

Fxxx, ,..., ==X B xCIxBii ∈:Π ⎡⎤ ∈ (12) tt12, ,..., tnJ()() 1 2 nccΠ CI()({ () ( ) J⎣⎦ () })

where, B =−∞(),,xx12 ×−∞( ) ×…×−∞(), xN . ■

The above approach will be discussed comprehensively in Chapter 3. To connect the Definition 1.3 with the first Definition 1.1 we can think of [ as the space of real function of a real variable DI(). We then have the following

Theorem 1.7 [Kolmogorov Theorem] : For every given family of finite–dimensional distributions satisfying conditions a), b), there exists a probability space ()Ω,,Uc and a stochastic process X ()t ;β , tI∈ defined on it that possesses the given distribution as finite – dimensional distributions. Proof : For the proof we refer to KLIMOV, G., ( Probability Theory and Mathematical Statistics. p.109) and ASH, R.B., (Probability and Measure Theory. p. 191). ■

In what follows we assume the existence of a probability space (Ω, Uc()Ω , ) . The stochastic process X :Ω→CI() is defined on it and the induced probability space is

()CI(),,Uc() CI () CI() .

2.1.2. Moments

As in the case of random variables, some of the most important properties of a s.p. are characterized by its moments, particularly those of the first and the second order. In the sequel, the existence of density functions shall be assumed.

The moments of the probability density function ft ( x) , are defined as the expectation n value EXtβ (()();β ). We have +∞ RnntEXt≡=β ;β n xfxdx (13) Xt()() ( ) ∫ () −∞

As special cases we have the first order moment

+∞ R1 tmtEXt==β ;β = xfxdx XX() ()() ( ) ∫ t() −∞ which is the mean value of Xt(); β at time t . Additionally the second order moment

+∞ R22tmtEXt==β ; β 2 = xfxdx (14) XX() ()() ( ) ∫ t() −∞

24 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

is the variance of Xt(); β at time t . Analogously we can define n–point moments. We have

nn n RtttEXtXtXtnn12⋅ ⋅⋅⋅ nm , ,...,≡⋅⋅⋅=β ;ββ12 ; ... ; βm XX... X()()()() 1 2 m() 1 2 m +∞ (15) =⋅⋅⋅xnn12x... xnm f() x , x ,..., x dx ... dx ∫ 1 2mttt12 , ,...,m 1 2 m 1 m −∞

A special case is the autocorrelation function,

β RXX () tt12,;;≡⋅ E() Xt ( 1β ) Xt( 2β ) (16)

One very important moment of a stochastic process X ()t ;β is the autocovariance function defined by

β CXX () tt12,;≡−⋅− E()⎣⎦⎣⎦⎡⎤⎡⎤ Xt ( 1ββ )()( mt 1 Xt 2 ; )() mt 2 (17)

The importance of the correlation functions or, equivalently, the covariance functions rests in part on the fact that their properties define the mean square properties of the stochastic processes associated with them. This will become clear in Chapter 4. In what follows, we note some of the important properties associated with these functions.

1. RXX()tt12,,= R XX () tt 21 for every tt12, ∈ I (18)

2 2. RXX()tt12,,,≤⋅ R XX () tt 11 R XX () tt 22 for every tt12, ∈ I (19)

3. RXX ()tt12, is positive–definite on I×I . (20)

2.1.3. Characteristic Functions

Following the definition of a characteristic function associated with a sequence of random variables, the characteristic functions of a s.p. X ()ttI;,β ∈ , are defined in an analogous way. There is, of course, a family of characteristic functions since X ()t ;β is characterized by a family of density functions. Again, we assume the existence of the density functions in what follows. The characteristic function φ u is defined in the following manner for the one dimensional t1 () case

+∞ φ uEe≡=β iux efxdx iux (21) tt() ()∫ () −∞

That is, φ1 (u) is the Fourier transform of f1 ( x) . By setting u = 0 we can note immediately that

+∞ φ ()01==f ()xdx (22) tt11∫ −∞

2.1. DEFINITION OF STOCHASTIC PROECESSES 25

The characteristic functional may be defined for n variables uu12, ,..., un analogously. Here

⎛⎞⎡⎤n φβuu, ,..., u≡⋅= Eβ exp i Xt ; u tt12, ,..., tn () 1 2 n⎜⎟⎢⎥∑ () k k ⎝⎠⎣⎦k=1 (23) +∞ +∞ +∞ ⎡⎤n =⋅⋅... expixuf() xxxdxdx , ,..., ... ∫∫∫ ⎢⎥∑ kk ttt12, ,...,n 1 2 n 1 n −∞ −∞ −∞ ⎣⎦k=1

Here are some of the basic properties of the characteristic function of a stochastic process:

1) φφuu, ,..., u ≤= 0,0,...,0 1 tt12, ,..., tnn() 1 2 n tt 12 , ,..., t ( )

2) φφ()uu=− ( ) tt12, ,..., tnn tt 12 , ,..., t 3) φφuu, ,...,0= uu , ,..., u tt12, ,..., tnn() 1 2 tt 12 , ,..., t− 1( 1 2 n− 1 ) 4) φ u is uniformly continuous for each u ∈ n tt12, ,..., tn () 5) φ ()u is a positive–definite function tt12, ,..., tn

The extension of φ to infinitely many variables involves the concept of a functional and will be considered in Chapter 3. There is an intimate connection between the moments and derivatives of φ with respect to u . Following the analysis we made for random variables at Section 1.3 of Chapter 1 we will have the similar result

++… ∂nn1 m nn12⋅ ⋅⋅⋅ nm 1 RtttXX... X() 1, 2 ,..., n=⋅++… φ t t… t () 0,0,… ,0 . (24) iuunn1 mm∂∂n1 … n12 n

26 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

2.2. Classification of Stochastic Process

In what follows we assume the existence of a probability space (Ω, Uc()Ω , ) . The stochastic process X :Ω→CI() is defined on it and the induced probability space is

()CI(),,Uc() CI () CI() .

We have seen that a s.p. X ()t ;β , tI∈ , is defined by a system of probability distribution functions. This system consists of

FxxxXtx, ,...,=< ,ββ ,..., Xtx , < tt12, ,..., tn () 1 2 nc ()( 1) 1∩∩( n) n

for every n and for every finite set {}tt12,,… , tn in the fixed index set I. In general, all probability distribution functions are needed to specify a stochastic process completely. It is not uncommon, however, that stochastic processes modelling physical phenomena are statistically simpler processes, simple in the sense that all the statistical information about a given process is contained in a relatively small number of probability distribution functions. The statistical complexity of a stochastic process is determined by the properties of its distribution functions and, in the analysis of stochastic processes, it is important to consider them in terms of these properties.

We consider in the remainder of this chapter classification of stochastic processes according to the properties of their distribution functions. Presented below are two types of classifications that are important to us in the analysis to follow, namely, classification based upon its memory and classification based upon the statistical regularity of a process over the index set I. We note that these two classifications are not mutually exclusive and, in fact, many physically important processes possess some properties of each.

2.2.1. Classification Based Upon Memory

In this classification, a s.p. X ()t ;β , tI∈ , is classified according to the manner in which its present state is dependent upon its past history. It is centred around one of the most important classes of stochastic processes, that is, the Markov processes.

2.2.1.a. Purely Stochastic Process – Generalized Stochastic Process

According to memory properties, the simplest s.p. is one without memory. They characterized by the following

Defintion 2.1 : A s.p. X ()t ;β , tI∈ is fully stochastic when the random variables Xt( 1 ;β) and Xt()2 ;β are independent for every tt12≠ . ■ From the above definition it is clear that the nth order distribution function can be written as product of first order distributions. 2.2. CLASSIFICATION OF STOCHASTIC PROECESSES 27

n FxxxFx, ,..., = (1) tt12, ,..., tni()() 1 2 n∏ t i i=1

The probability structure of the fully stochastic process is characterized from the first order distribution only. Although mathematically simple, a continuous-parameter purely stochastic process is not physically realizable because it implies absolute independence between the present and the past no matter how closely they are spaced. In physical situations we certainly expect Xt( 1 ;β) to be dependent on Xt()2 ;β when t1 , is sufficiently close to t.2 In practical applications, therefore, purely stochastic processes with a continuous parameter are considered to be limiting processes. In order to obtain a mathematic satisfactory theory of the process just described one has to introduce the concept of generalized s.p. ∞ Let CIc () be the space of all infinitely differentiable functions ϕ()tI: → vanishing identically outside a finite closed interval; it is a space of test functions. Let m be a of random variables defined on (Ω, Uc(Ω), ) having finite second moment.

∞ Defintion 2.2 : A generalized s.p. on I is a continuous linear mapping Φ from CIc () to m ; the value of a generalized stochastic process Φ at ϕ will be denoted by {}ϕ,Φ or Φ(ϕ) . ■

Example 2.3 : Let X ()t ;β , tI∈ be a s.p. with finite second moment (a mapping from I to ∞ m ). As we shall see in Chapter 4 , for a very wide class of processes and for ϕ()tCI∈ c ( ) there exists the following integral

ϕϕϕβ,;Φ=Φ= tXt dt (2) {}()()()XX ∫ S where S is an arbitrary compact set in I ( X ()t ;β is a locally integrable stochastic process on I ). The above mapping is linear and continuous. Therefore, every locally integrable s.p. on I determines a generalized s.p. in the form presented above. ■

One of the important advantages of a generalized stochastic process is the fact that its derivative always exists and is itself a generalized s.p.

∞ Defintion 2.4 : The derivative Φ′ with respect to t of a generalized s.p. Φ in CIc () is defined by means of the formula

⎪⎪⎧⎫dϕ ∞ {}ϕ ,,Φ=−′ ⎨⎬ Φ for all ϕ()tCI∈ c ( ) ■ (3) ⎩⎭⎪⎪dt

⎪⎪⎧⎫dϕ Remark 2.5 : It can be verified that the mapping ϕ →Φ⎨⎬⎪⎪, is a continuous linear mapping ⎩⎭⎪⎪dt ∞ (α) from CTc () to m . The derivative Φ for any given α ≥ 1 are given by

⎧⎫α ()α α ⎪⎪d ϕ {}ϕ ,1,Φ=−()⎨⎬α Φ ■ (4) ⎩⎭⎪⎪dt Remark 2.6 : It is clear from Example 2.3 that the appropriate definition of ϕ ,Φ ′ is {}X

28 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

ϕϕβ,;Φ=−′ ′ tXt dt (5) {}X ∫ () ( ) S From the above it follows that an ordinary process X ()t ;β may also be regarded as always differentiable, provided its derivative be a generalized process instead of an ordinary process. ■ From the standpoint of correlation theory, a generelazed s.p. Φ or Φ(ϕ) is characterized by its mean

β mEΦ ()ϕϕ=Φ() () (6) and correlation operator

β ∞ CE()ϕψ,,=ΦΦ() ()() ϕ ψ ϕψ ∈ CIc () (7) which is assumed to be linear and continuous with respect to ϕ and ψ .

A typical example of a generalized stochastic process is Gaussian white noise ξ .

Definition 2.7 : A Gaussian white noise ξ or ξϕ() is a generalized s.p. such that there exist a Wt();β tI∈ satisfying ξβ()tWt;;= ′ () β, i.e.

⎪⎪⎧⎫dϕ ∞ {}ϕξ,,==−{} ϕWW′ ⎨⎬ , for all ϕ()tCI∈ c ( ) ■ (8) ⎩⎭⎪⎪dt

∞ Note that for every ϕ()tCI∈ c ( ), {}ϕξ, is a Gaussian r.v., and that for every finite number of ∞ functions ϕϕ12,,,… ϕn in CIc (), the random variables {}ϕξi ,;1≤≤in are jointly Gaussian. This is the reason why ξ is called a Gaussian white noise.

Theorem 2.8 : If ξ is a white Gaussian noise, then

β ∞ E ()ξϕ()= 0 for all ϕ()tCI∈ c ( ) (9) and CttdtCIϕψ,,,=∈ ϕ ψ ϕψ ∞ (10) ξ ( )∫ () ()c ( ) T

Proof : ⎛⎞ ββ β⎜ ddϕϕ⎟ β EE()ξϕ()==−=−−=(){} ϕξ,0 EWtdtEWtdt⎜ () ⎟ ()() ⎜∫∫⎟ ⎝⎠TTdt dt since EWtβ ( ( ))= 0 for all tI∈ and ϕ is arbitrary. To show the second equality, let ∞ ϕψ, ∈ CIc (). Then ⎛⎞ ββ⎜ ddϕψ⎟ CEξ ()ϕψ,,,=⋅=−⋅−=() {}{} ϕξ ψξ E⎜ WtdtWtdt()11() 2 2⎟ ⎜∫∫⎟ ⎝⎠TTdt dt ddϕψ dd ϕψ =⋅=Eβ W t W t dtdtmin t , t dtdt ∫∫()()1 () 2 12 ∫∫ () 12 12 TTdt12 dt TT dt 12 dt

After integrating by parts we obtain the desired equality. ■

2.2. CLASSIFICATION OF STOCHASTIC PROECESSES 29

Remark 2.9 : It is worth noting that the formula (10) can be put in the form

CtsttdsdtCIϕψ,,,=− δ ϕ ψ ϕψ ∈∞ (11) ξ ( )∫∫ ( ) () ()c ( ) TT which implies that the correlation function of the Gaussian white noise is the generalized function δ()ts− . ■

2.1.2.b. Markov Stochastic Process

The Markov s.p., is a subset of the wide class of s.p. which characterized only from the second order distribution function. More specifically, for the Markov s.p., the random variables defined for all the values of t > τ are independent from the random variables defined for t < τ if we know the value of the stochastic process at time τ . The above property is expressed at the following

Defintion 2.10 : A s.p. X ()t ;β , tI∈ is Markov, if for every n and every {}tt12<<… < tn in I we have

Fxx|,;,,,, t x t… xt= Fxx |, t (12) tnnnnnn ()−−11 − 2 − 2 11 tnnnn ( −− 11) or equivalently,

XX ccU()Bx|,;,,,,nn−−11 t x n − 2 t n − 2… xt 11=∈( Bx |, nn −− 11 t) , B () (13) ΠΠ()ttn ()n

where Π ()tn is the projection operator Π (txnn) ⎣⎦⎡⎤(i) = xt( ) for every tn .(See Example 1.6. of Section 2.1) ■

Remark 2.11 : The equivalent definition, presented above, can be used to generalize the concept of Markov s.p. to the case of a general Hilbert space, whose elements, are defined on some ordered set. ■

As we can easily see, the conditional distribution of Xt()n ;β for given values of

Xt()()12;,ββ Xt ;, …,;Xt( n−1 β) depends only at the most recent known value Xt()n−1 ;β . The above definition implies that the basic characteristics of a Markov Process are described by a function

FxtysFxysMt()(),; ,= | , (14)

The function FxtysM (),; , is called a transition probability. It gives the probability of a transition of a process from a state y at time s to one of the states for which {X ()tx;β < } at time t . We can easily see that for fixed tsy,, the transition probability FM ()xt,; ys , is a probability distribution function.

Proposition 2.12 : The probability structure of a Markov s.p. is completely characterized by the second order probability distribution functions.

30 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

Proof: To see that a Markov process is completely specified by its first and second distribution functions, we start with the general relation Fxxx, ,..., = tt12, ,..., tn () 1 2 n

=⋅⋅Fxxtxttnnnnn( |,;,,,,−−11 − 2 − 2…… xtFx 11) t( n − 1 |,,,, xt nn − 2 − 2 xt 11) n n−1 (15) ⋅⋅… FxxtFx|, ⋅ tt21()()211 1

Using the definition of the Markov s.p. we immediately see that for n ≥ 3 and

{}tt12<<… < tn n−1 FxxxFxFxxt, ,...,=⋅ | , (16) tt12, ,..., tni()()() 1 2 n t 1 1∏ t+ 1 i+ 1 i i i=1 But, Fxx(), Fxxt()|, = ttii+1 ,1 i+ i (17) tiiii+1 +1 Fx tii () Consequently, n−1 Fxx, ∏ ttii+1 ,1() i+ i Fxxx, ,..., = i=1 ■ (18) tt12, ,..., tn () 1 2 n n−1 Fx ∏ tii () i=2

Proposition 2.13 : Let ttt123<<. For the transition probability density function the following Smoluchowski-Chapman-Kolmongorov (SCK) holds,

+∞ f xt,; xt ,= f xt ,; xt , f xt ,; xt , dx (19) MMM()()()33 11∫ 33 22 22 11 2 −∞ or equivalently, +∞ Fxtxt,; ,= FxtxtdFxtxt ,; , ,; , (20) MMM()()()33 11∫ 33 22 22 11 −∞

Proof: Consider the general relation

+∞ f ()xx,,,= f ( xxx ) dx (21) tt13,13∫ tt 123 ,,1232 t −∞

Let ttt123<<. Then using the relation

f xxx,,=⋅ fxt ,|,;, xt xt fxt ,|, xt f x tt123,, t()( 1 2 3 1 1 2 2 3 3 )( 2 2 3 3) t3 ( 3) and equations (15) we clearly have

f xxx,,=⋅⋅ fxt ,|, xt fxt ,|, xt f x tt123,, t()() 1 2 3 1 1 2 2( 2 2 3 3) t3 ( 3)

Then making use of equation (21) we have the desired equation. ■

In the analysis of Markov Processes and in modeling of real random phenomena we frequently deal with so called stationary transition characterizing a homogeneous Markov s.p.

2.2. CLASSIFICATION OF STOCHASTIC PROECESSES 31

Definition 2.14 : A Markov s.p. X ()t ;β , tI∈ is said to be homogeneous (with respect to time) if for arbitrary sI∈ and tI∈ , st< , the transition probability depends only on the difference τ =t− s, that is

FMHM()()xt,; ys ,= F x ,τ ; y ■ (22)

The function FHM ()xy,;τ determines the probability of transition from a state y at time s to one of the states for which {X ()tx; β < } at time t . This probability does not depends on the position of s , but only from the difference ts− . In this case the Smoluchowski-Chapman- Kolmongorov equation takes the form

+∞ Fx,;ττ+= x Fx ,; τ xdFx ,; τ x (23) HM()()()11 2 3∫ HM 11 2 HM 2 2 3 −∞ The SCK equation has important implications in the theory of Markov s.p.; among others, as it can be shown (SOBCZYK, K., Stochastic Differential Equations. p. 31), it implies the existence of a semi – group of operators associated with a homogeneous Markov s.p..

Preposition 2.15 : Let the homogeneous Markov s.p. X ()t ;β , tI∈ with transition probability

FHM ()xy,;τ . For every bounded, measurable, function f defined on the state space

R ⊂ (the set of possible values of the s.p.) of the process we define the operator Tt acting on the functions f as

Tf⎡⎤i = fxdFxty,; (24) tHM⎣⎦()∫ ( ) ( ) R then,

a) Operators Tt are linear, positive and continuous.

b) Operators Tt form a semi–group, i.e.

TTTTTstts+ == t s s t ,0 >. (25)

Proof: For the proof of a) we refer to (SOBCZYK, K., Stochastic Differential Equations. p. 31). Part b) can be proved using the form (23) of Smoluchowski-Chapman-Kolmongorov equation. ■

2.1.3.c. Processes with Independent Increments

Mathematical models of numerous real phenomena can be described by stochastic processes whose increments are independent random variables.

Definition 2.16 : A s.p. X ()t ;β , tI∈ is said to be an independent–increment process, if for every n and every {}tt12<<… < tn in I , the random variables,

Xt()()12;,ββ Xt ;−− Xt () 1 ;,, β… Xt ()nn ; β Xt (− 1 ; β ) are independent. ■

32 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

The definition above implies that all finite–dimensional distributions of the process X ()t ;β are completely determined by the distribution of the random variable X ()t ;β for each t and the distributions of the increments Xt()()21;;ββ− Xt for all possible values of tt12, , such that tt12< . However, as we shall show at the next preposition, the distributions of the of the increments Xt()()21;;ββ− Xt can be expressed in terms of the distributions of X ()t ;β . So, we have the following,

Remark 2.17 : For the special case of an independent increment process, where X ()0;β = 0 almost surely we will have Xt();;;;0;βββββ=−+−() Xt () Xs ()() Xs() X () where ()Xt();;ββ− Xs () and ()Xs();0;ββ− X () are independent random variables. Consequently stochastic process X ()t ;β verifies the Markov property. ■ Preposition 2.18 : Let the independent–increment s.p. X ()t ;β , tI∈ . Then all the finite– dimensional distributions of the process X ()t ;β can be expressed in terms of the one dimensional distribution of the random variable X ()t ;β for every fixed t.

Proof : We will prove the above assertion in two steps. First we have for the general finite dimensional characteristic function ⎛⎞⎡ n ⎤ φβuu, ,..., u≡⋅= Eβ exp i Xt ; u tt12, ,..., tn () 1 2 n⎜⎟⎢ ∑ () k k⎥ ⎝⎠⎣ k =1 ⎦ ⎛ ⎡ ⎛⎞nn ⎛⎞ =+×−++EiuXtiuXtXtβ exp ;βββ⎡⎤ ; ; ... ⎜ ⎢ ⎜⎟∑∑ii()121 ⎜⎟⎣⎦ ()() ⎝ ⎣ ⎝⎠ii==12 ⎝⎠ n ⎤ ⎞ +×i∑ uin⎡⎤⎣⎦ Xt()(;;ββ − Xt n−1 )⎥ ⎟ = in= ⎦ ⎠ =+++⋅++⋅⋅φφuu... u u ... u… φ u (26) tnttnttn1121()12 ,( 2)()nn− ,

Let us now take two arbitrary time instants tt12< . It is clear that Xt;;ββ=+ Xt⎡ Xt ;; ββ − Xt ⎤ , i.e. Xt;β is a sum of two independent random ()21 ()⎣ ()() 21⎦ ()2 variables. Hence the following relation holds for the characteristic functions,

φ ()u φ ()u = t1 (27) tt12, φ u t2 () for each tt12< . Using the eq.(26), we have φ (uu++... ) φ ()u φφ()()uu, ,..., u=+++⋅ u u ... u tn1 2 ⋅⋅… tnn−1 ■ (28) tt12, ,..., tn 1 2 n t 1 1 2 n φφuu++... u tntn2 ()2 n () Remark 2.19 : It should be noticed that the one dimensional distribution of the independent– increment s.p. cannot be given arbitrary. The one dimensional characteristic function has to be such that φ u given by (27) is also a characteristic function. ■ tt12, ( ) An important class of independent–increment s.p. is formed by the processes with orthogonal increments. The Riemann–Stieljes integrals with respect to process with orthogonal increments 2.2. CLASSIFICATION OF STOCHASTIC PROECESSES 33 are particularly important in the study of stationary stochastic processes. Moreover, as we shall prove at the sequel every stochastic process has orthogonal increments if and only if is a martingale process (Preposition 2.39).

Defintion 2.20 : A s.p. X ()t ;β , tI∈ is said to be an orthogonal increment process, if for every tsuv<≤< in I , β ⎡⎤2 EXt()⎣⎦();;ββ−<∞ Xs () (29) and β ⎡⎤⎡⎤ EXt()⎣⎦⎣⎦();;;;0ββββ−⋅− Xs () Xu ( ) Xv ( ) = ■ (30) Remark 2.21 : From the above definition we can easily see that an independent–increment s.p. that fulfils eq. (29) and has constant mean value, i.e. mtX ()= const. is also an orthogonal increment process. ■

We now define the function ⎧ 2 ⎪ EXtβ ⎡⎤;;,ββ−≥ Xt tt ⎪ ()⎣⎦() (00 ) Ftt(); = ⎨⎪ (31) 0 ⎪ 2 ⎪−−EXtβ ⎡⎤();;,ββ Xt ( ) tt < ⎩⎪ ()⎣⎦00

Clearly, we have Ftt();;00=− Ft () t for every tt, 0 ∈ I. Theorem 2.22 : Let the orthogonal increment s.p. X ()t ;β , tI∈ . Then for the variance of the increment Xt()12;;ββ− Xt () for every tt12< with tt12, ∈ I the following relation holds

2 =+β ⎡⎤ββ − ∈ Ftt();;12 Ftt () E()⎣⎦ Xt ( 1 ; ) Xt ( 2 ;, ) t I (32)

Proof : Without any loss of generality we can assume tt12< . Under this assumption we have three cases. Indeed, consider the new constant t1 . This constant can be choose such that ttt10<< or ttt01<<. We examine the first case. (For the sequel we will denote F (tt; i ) as

Fti ()) Case I) :

t t t 1 2 2 =−β ⎡⎤ββ − = Ft11() E()⎣⎦ Xt (;; ) Xt ( )

2 =−β ⎡⎤ββ − + ββ − = EXt()⎣⎦()()()();; Xt122 Xt ;; Xt

22 =−Eββ⎡⎤⎡⎤ Xt;;ββ − Xt − E Xt ;; ββ − Xt − ()⎣⎦⎣⎦() (221 ) ()()()

−−⋅−β ⎡⎤⎡⎤ββββ 2;EXt()⎣⎦⎣⎦()()()() Xt221 ; Xt ; Xt ; 22 =−ββ⎡⎤⎡⎤ββ − + ββ − = F221() t E()⎣⎦⎣⎦ Xt (;;2;; ) Xt ( ) E() Xt()() 21 Xt

2 =+Ft Eβ ⎡⎤ Xt;;ββ − Xt 221()()⎣⎦ ( ) ( )

34 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

Case II) :

t t t 1 2

2 =−=β ⎡⎤ββ Ft11() E()⎣⎦ Xt (;; ) Xt ( )

2 =−+−=β ⎡⎤ββ ββ EXt()⎣⎦()()()();; Xt122 Xt ;; Xt

22 =−ββ⎡⎤⎡⎤ββ ββ −+ E()⎣⎦⎣⎦ Xt();; Xt (221 ) E() Xt()() ;; Xt +−⋅−β ⎡⎤⎡⎤ββββ 2;EXt()⎣⎦⎣⎦()()()() Xt221 ; Xt ; Xt ;

22 =−ββ⎡⎤⎡⎤ββ ββ −+ E()⎣⎦⎣⎦ Xt();; Xt (221 ) E() Xt()() ;; Xt −−⋅−2;EXtXtXtXtβ ⎡⎤⎡⎤ββββ ;; ; ()⎣⎦⎣⎦() (22 ) () ( ) 22 =−EXtβ ⎡⎤⎡⎤;;ββ−+Xt Eβ Xt ;; ββ −= Xt ()⎣⎦⎣⎦() (221 ) ()()()

2 =+β ⎡⎤ββ − Ft221() E()⎣⎦ Xt (;; ) Xt ( ) Case III) :

t t t 1 2

2 Ft=−= Eβ ⎡⎤ Xt;;ββ Xt 11()()⎣⎦ ( ) ( )

2 =−+−=β ⎡⎤ββ ββ EXt()⎣⎦()()()();; Xt122 Xt ;; Xt

22 =−EXtββ⎡⎤⎡⎤;;ββ Xt EXt ;; ββ −+ Xt ()⎣⎦⎣⎦() (221 ) ()()() +−⋅−2;EXtβ ⎡⎤⎡⎤ββββ Xt ; Xt ; Xt ; ()⎣⎦⎣⎦() (221 ) ( ) ( ) 2 =+Ft Eβ ⎡⎤ Xt;;ββ − Xt 221()()⎣⎦ ( ) ( ) We thus have complete the proof. ■

Hence, we see that the change of the parameter t0 change the function F ()tt; 0 only by a positive constant. For the sequel we will denote Ftt(); 0 as Ft( ) and we will call it as the characteristic of X ()t ;β . From the above theorem we can easily see that the variance of the increment Xt()12;;ββ− Xt () can be described by

2 β ⎡⎤ββ−=−<∈ EXt()⎣⎦()12;; Xt () FtFt() 2112 () , ttI (33)

Thus, for every orthogonal increment process we have 2.2. CLASSIFICATION OF STOCHASTIC PROECESSES 35

Preposition 2.23 [Levy property] : Let the s.p. X ()t ;β tI∈ with orthogonal increments and

at=<<12 t… <= tn b an arbitrary partition of the interval [ab, ] and Δ =−max()ttii+1 . Then, n−1 2 ⎡⎤nn l.i.m.X ()()tXtFbFaii+1 ;ββ−=− ; () () ■ (34) Δ→0 ∑ ⎣⎦⎢⎥ n→∞ i=0 Now we are going to describe some specific cases of s.p.’s described above. The most important examples of independent–increment s.p.’s are the Wiener process and the Poisson process. For a complete study of processes with independent increments we refer to PUGACHEV, V.S. & SINITSYN, (Stochastic Differential Systems).

a) Wiener Process

Definition 2.24 : A s.p. Wt();β tI∈ is called a Wiener process or Brownian process if: X a) ccΠ ()tCI=0 (){}0001= ()(){xCIx()i ∈∈= (): () {}}

b) for every n and every {}tt12<<… < tn in T , the increments Wt()()21;;,,ββ− Wt …

Wt()(nn;;ββ− Wt−1 ) are independent c) for arbitrary t and h > 0 the increment Wt()()+− h;;ββ Wt has a Gaussian distribution with

EWthβ ()()()+−;;ββμ Wt = h β ⎡⎤2 2 EWth()⎣⎦()()+−;;ββσ Wt = h where μ a real constant, called drift coefficient and σ 2 a positive constant, called variance. ■ Without restriction one can assume that μ = 0 and σ 2 =1. In what follows we consider this case (standard Wiener process).

Preposition 2.25 : The standard Wiener s.p. Wt();β tI∈ has the following properties: a) Wt();β is a centred, second–order, Gaussian s.p., i.e.

β β ⎡⎤2 EWt()();0β = and EWt()⎣ ();β ⎦ <∞ (35)

This process is not stationary and its covariance function is written as

β CWW () tt12,;;min,,,=⋅= E() Wt ()() 1ββ Wt 2() tt 12 tt 12 ∈ I (36)

b) Wt();β is a Markov s.p. whose system of transition probabilities is homogeneous and for τ > 0 , is written as 2 z 11⎪⎧ ()yx− ⎪⎫ FxzHM (),;τ =−⋅ exp⎨ ⎬ dy (37) ∫ 2 τ −∞ 2πτ ⎩⎭⎪ ⎪ c) Wt();β is mean–square continuous on I . d) The joint probability density function has the form

36 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

2 ⎧⎫n n W ⎪⎪11()xxii− −1 fxxxtt,,,… t() 1,,, 2… n =−⋅ exp⎨⎬ (38) 12 n ∑ 2 tt− ∏ ⎩⎭⎪⎪i=1 ii−1 i=1 2π ()ttii− −1

Proof : We only give some elements of the proof herein. For more details we refer to SOIZE, C. (Mathematical Methods in Signal Analysis). For the proof of point a) we can write for t > 0 ,

Wt();;0;0;ββββ=−+() Wt () W ( ) W()

We then use Definition 2.24. Point b) can be proved by direct application of Remark 2.17 and using the Gaussian structure proved at point a). The proof of point c) is obtained by calculating the mean square norm of the difference, i.e. by direct application of m.s.-continuous definition.

For point d) we notice that the vector ()Wt()1 ;,,ββ… Wt ()n ; is a non-degenerate linear transformation of the Gaussian vector ()Wt()10;;,,;;ββ−− Wt ()… Wt ()nn β Wt (− 1 β ). For the last vector the density is given by

n 2 n ΔW ⎧⎫11xi fxxxtt,,,… t() 1,,, 2… n =−⋅ exp⎨⎬ (39) 12 n ∑ 2 tt− ∏ ⎩⎭i=1 ii−1 i=1 2π ()ttii− −1

Hence after some calculus we get the desired result. ■

The fundamental properties of the realizations of a Wiener s.p. are summarized in the following

Theorem 2.26 : Let the Wiener s.p. Wt();β tI∈ . Almost all realizations (sample functions) of Wt();β , are continuous, nowhere differentiable and have unbounded variation in every finite interval.

Proof : The continuity follows from the Kolmogorov criterion which will be formulated in Section 2/Chapter 4 (see Example 2.8/Chapter 4). The non–differentiability for fixed t can be made clear as follows: the distribution of the difference quotient

1 ⎡⎤Wt()()+− h;;ββ Wt h ⎣⎦

⎛⎞1 is the Gaussian distribution N ⎜0, ⎟ which diverges as h → 0; therefore this quotient can not ⎝⎠⎜ h⎟ converge with positive probability to a finite random variable in any probabilistic sense. The last assertion of the theorem follows from Preposition 2.23 and some additional considerations connected with almost sure convergence. ■

b) Poisson process

Definition 2.27. A s.p. Nt();β tI∈ is called a non–homogeneous Poisson process if

a) the s.p. Nt();β has independent increments

N b) ccΠ ()tDI=0 (){}0001= ()(){NDIN()i ∈∈= (): () {}} 2.2. CLASSIFICATION OF STOCHASTIC PROECESSES 37

c) for arbitrary tt12, the difference Nt( 12;;ββ)− Nt( ) have a Poisson distribution, that is there exist a function λ ()t that for each integer number k > 0

N ccΠ ()tt, ()k=∈−∈= DI()(){ N()i DI( ): Nt( 12) Nt( ) { k}} 12 k ⎡⎤λλ()tt− () = ⎣⎦12e−⎡⎣⎦λλ()tt12 − ( ) ⎤ k!

If λνν()tt=>,0 then the process Nt();β is called a homogeneous Poisson process. ■

It is clear that the processes Nt();β for each tI∈ are random variables with Poisson distribution

k ⎡⎤λ t N ⎣⎦() −λ()t ccΠ tDI()kNDINtke=∈∈={}()i ( ): () {} (40) () ()()k!

Preposition 2.28 : Let Nt();β be a Poisson process with mean value λ ()t . Then Nt();β is a Markov process with transition probability

k ⎡⎤λλts− N ⎣⎦() () −−()λλ()ts ( ) ccΠ ts, ()kNtNsk=−∈= DI {}() ( ) {} e (41) () ()()k!

For the case when λ (tt) = we have a homogeneous Poisson process and the homogenous transition probability has the form

k N ()ts− −ts+ ccΠ ts, ()kNtNske=−∈= DI {}() ( ) {} (42) () ()()k!

Proof : Direct application of Remark 2.17 and equation (14). ■

The following assertions characterize the basic properties of the Poisson process.

Theorem 2.29 : For an homogeneous Poisson process with intensity ν , the interarrival times tt12,,… , tn are independent and identically distributed random variables with the common distribution being exponential with parameter ν . ■

Proof : For the proof we refer to SNYDER, D.L., (Random Point Processes). ■

Remark 2.30 : It is however, worthy adding that interarrival times for an inhomogeneous Poisson process are not independent. ■

The fundamental properties of the realizations of a Poisson s.p. are summarized in the following

Theorem 2.31 : Almost all sample functions of the Poisson process are non-decreasing with integer increments; they increase by jumps equal to one and the number of jumps in every finite interval is finite. ■

Proof : For the proof we refer to SNYDER, D.L., (Random Point Processes). ■

38 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

2.1.3.d. Martingales

An important class of processes is formed by martingales. We first recall the definition of a current of σ − algebras of a probability space. Let I denotes the set of real numbers corresponding to the times at which the stochastic experiments were carried out.

Definition 2.32 : Let the probability space (Ω, Uc(Ω), ) . A monotonically decreasing family of σ − algebras U t()Ω , tI∈ with

a) UUt()Ω⊂ () Ω

b) ()Ω⊂ () Ω if tt< UUtt12 12 is called a current of σ −algebras. ■

Definition 2.33 : Let the s.p. X ()t ;β , tI∈ , defined on the probability space

()Ω, Uc()Ω , , which is equipped with a current of σ − algebras U t()Ω , tI∈ . The s.p.

X ()t ;β , tI∈ is adapted to the current of σ − algebras U t()Ω , tI∈ iff the random variable

X ()t ;β is U t()Ω - measurable for every tI∈ . ■

Remark 2.34 : It follows from the definition that U t()Ω always contains the σ − algebra generated by the random variables {}X ()sst;,β < . ■

Remark 2.35 : For every real valued s.p. X ()t ;β , tI∈ , one current of σ − algebras U t()Ω is the following,

X −1 UUt ()Ω=σ(){}XsAs (;: ) ∈ [ 0,, tA ] ∈ () (43) where σ()A denotes the generated σ − algebra by the set A . The inverse sign is for the X random argument A . As we can understand U t ()Ω is representing a σ −algebra of the X physical events, as they have been evolved, until time t , i.e. U t (Ω) characterizes a collection of events representing the known information at time t . The above remark can be extended for the case of a vector or complex valued s.p., with appropriate change to the Borel sets U () . ■

Definition 2.36 : Let the s.p. X ()t ;β , tI∈ . We assume that X ()t ;β is U t()Ω - measurable. The above s.p. is called a maringale iff

ββ EXt()();|ββUUss ()Ω= EXs()() ;| () Ω= Xs() (44) for all st , ∈ T, and st≤ . ■

Remark 2.37 : Often the term martingale is referred to processes for which relation (44) holds for the family of σ − algebras generated by the processes X ()t ;β itself (Remark 2.34.), what means that the history of the process X ()t ;β prior to the instant t is chosen as condition. ■

Example 2.38 : A Wiener s.p. Wt();β tI∈ , is a martingale, since

2.2. CLASSIFICATION OF STOCHASTIC PROECESSES 39

Eββ Wt;|ββββΩ= E Ws ; +⎡ Wt ; − Ws ;⎤ | Ω= ()()UUss () ()()⎣ () ()⎦ () =+Ws Eβ ⎡ Wt;;|ββ − Ws ⎤ Ω= ()()⎣ ( ) ( )⎦ U s ( ) β ⎡ ⎤ =+Ws() E()⎣ Wt (;;ββ ) − Ws ( )⎦ = Ws() since the random variable Wt();;ββ− Ws () is independent of all events of σ − algebra

U s()Ω . ■

Preposition 2.39 : The stochastic process X ()t ;β , tI∈ is a square integrable martingale iff is an orthogonal increment process. Thus, every martingale can be described by a characteristic and satisfies the Levy property.

Proof : If X ()t ;β , tI∈ is a square integrable martingale then we clearly have for every ts, ∈ I

β ⎡⎤ EXt()⎣⎦();;0ββ−= Xs ()

Moreover for any disjoint sets [tt12, ]⊂ I and []ss12, ⊂ I we have

EXtβ ⎡⎤⎡⎤;;;βββ−⋅− Xt Xs Xs ;0 β =. ()⎣⎦⎣⎦()121 ()( ) ( 2 )

Hence X ()t ;β , tI∈ is an orthogonal increment process. If X ()t ;β , tI∈ is an orthogonal increment process then the independence of increments implies that for ts, ∈ I with st< , we have the independence of U s()Ω and X ()tXs;;ββ− (). (See Theorem 13.10 of YEH, J., Martingales and Stochastic Analysis). Thus

EXtββ⎡⎤⎡⎤;;|ββ−Ω=−= Xs EXt ;;0 ββ Xs ()⎣⎦⎣⎦() ()U s () ()() ()

Then

ββ⎡⎤ β EXt()();;;|;|βββ=− E() Xt() Xs ()UUss () Ω+Ω EXs()() β () ⎣⎦ = Xs();β

Therefore X ()t ;β is a martingale. For more details on the proof of the above theorem we refer to PROHOROV, YU.V. & ROZANOV, YU.A., (Probability Theory. p.137) and YEH, J. (Martingales and Stochastic Analysis. p.86). ■

Preposition 2.40 : Let X ()t ;β , tI∈ be a martingale s.p, U t()Ω - measurable. Then some basic properties, it possesses, are as follows: a) Sample functions of separable martingales have no discontinuities of the second kind; so, they have at worst jumps.

β b) EXt()();β , tI∈ is constant. So, for every tt12, ∈ I we have

β EXt()()()21;;0ββ−= Xt .

40 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

c) If Yt();β is another martingale with respect to the same family U t()Ω , then

aX12() t;;ββ+ aY () t (where aa12, ∈ ) is also a martingale, U t()Ω - measurable.

d) Moreover if Ψβ( ) is a U t()Ω - measurable random variable, then for every tt12, ∈ I, greater than t we have

EXtXtβ Ψβ⋅−⎡⎤;;0 β β = ()()⎣⎦ (21 ) ( )

Proof : The first assertions follows from Chentsov Theorem that it will be formulated in Theorem 2.9 of Chapter 4. The next two assertions follow directly from the definition of martingales. The proof of assertion d) can be found at PROHOROV, YU.V. & ROZANOV, YU.A., (Probability Theory. p.137). ■

2.2. CLASSIFICATION OF STOCHASTIC PROECESSES 41

2.2.2. Classification Based Upon Regularity

In terms of statistical regularities, stochastic processes can be grouped into two classes: stationary stochastic processes and nonstationary stochastic processes. The nth probability distribution function, F xx, ,..., x of a s.p. X t ;β , tI∈ , is in tt12, ,..., tn ( 1 2 n) () general a function of x12,,,xx… n and tt12,,… , tn . A nonstationary s.p. is one whose distribution functions depend upon the values of time parameters explicitly. The statistical behaviour of a nonstationary s.p. thus depends upon the absolute origin of time. Clearly, most stochastic processes modelling physical phenomena are nonstationary. In particular, all physical processes having a transient period or certain damping characteristics are of the nonstationary type. On the other hand, many stochastic processes occurring in nature have the property that their statistical behaviour does not vary significantly with respect to their parameters. The surface of the sea in spatial and time coordinates, noise in time in electric circuits under steady state operation have the appearance that their fluctuations as functions of time or spatial positions stay roughly the same. Because of powerful mathematical tools which exist for treating stationary s.p.'s, this class of s.p.'s is of great practical importance, and we consider it in some detail in the following sections. More exhaustive accounts of stationary stochastic processes can be found in the work of PUGACHEV, V.S. & SINITSYN , Stochastic Differential Systems.

2.2.2.a. Stationary and Wide-Sense Stationary Processes

Definition 2.41. A s.p. X ()t ;β , tI∈ , is said to be stationary or strictly stationary if its collection of probability distributions stay invariant under an arbitrary translation of the time parameter, that is, for each n and for an arbitrary τ,

F xxxF, ,...,=+∈= xxxt , ,..., ,τ Ij , 1,2,… , n ■ (45) tt12, ,..., tnn() 1 2 n t 1++ττ , t 2 ,..., t + τ( 1 2 n) j

Let τ =−t1 , in the above. We see that the probability distributions depend upon the time parameters only through their differences. In other words, the statistical properties of a stationary s.p. are independent of the absolute time origin.

We can easily see that a stationary s.p. X ()t ;β , tI∈ , possesses the following important properties for its moments, if they exist. Since the first distribution function Fxt () is not a function of t, we have

β EXt()();β == mX const.

Since

F xx,,= F xx tt12,120,12() t 2− t 1 () we have

β EXt()()(;;ββ Xt+=τ )CXX ()τ

In view of the general relation (18), it follows that

42 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

CXX()τ =−C XX (τ ) (46)

The correlation function of a stationary stochastic process is thus an even function of τ. Properties of higher moments can also be easily derived.

Given a physical problem, it is often quite difficult to ascertain whether the stationary property holds since Eq. (45) must hold for all n. For practical purposes, we are often interested in a wider class of stationary stochastic processes.

Definition 2.42. A s.p. X ()t ;β , tI∈ , is said to be wide-sense stationary if

β EXt()();const.β == mX <∞ (47) and

2 βββββ<∞ ⋅ = − E()() Xt();,;; E() Xt()()12 Xt CXX () t 21 t ■ (48)

A wide-sense stationary s.p. is sometimes called second-order stationary, weakly stationary, or covariance stationary. It is clear that a strictly stationary process whose second moment is finite is also wide-sense stationary, but the converse is not true in general. An important exception is the Gaussian s.p. As we shall see, a Gaussian stochastic process is completely specified by its means and covariance functions. Hence, a wide-sense stationary Gaussian stochastic process is also strictly stationary.

Example 2.43. Consider a s.p. X ()t ;β , tI∈ , defined by

Xt();cosβωΦ=+ a ( t ) (49) where a and ω are real constants and Φ is a r.v. uniformly distributed in the interval[]0, 2π . It is a simplified version of the Pierson Longuet-Higgins model discussed in Example 1.2 of Chapter 1. We easily obtain a 2π EXtβ ()();cos0βωφφ=+=() t d π ∫ 2 0 1 CttaE(),=++=−22β () cos (ωφ t )( cos ωφ t ) a cos() ω() tt XX 12 1 22 1 2 Equation (49) is thus wide-sense stationary. ■

2.2.2.b. Spectral Densities and Correlation Functions

We restrict ourselves here to the study of wide-sense stationary stochastic processes and develop further useful concepts in the analysis of this class of processes. It has been noted that the covariance or correlation function of a stochastic process plays a central role. Let us first develop a number of additional properties associated with the correlation function

β CXX ()τ =+⋅EXt() (τ ;;ββ )Xt ( ) of a wide-sense stationary s.p. X ()t ;β , tI∈ , 2.2. CLASSIFICATION OF STOCHASTIC PROECESSES 43

1) CXX ()τ is an even function of τ, that is

CXX()τ =−C XX (τ ) This has been derived in Section 5.2.1 Eq. (46).

2) CXX ()τ always exists and is finite. Furthermore 2 ≤=β β CXX()τ C0 XX () EXt()() (; ) (50) This result is a direct consequence of Eq. (3.19).

3) If CXX (τ) is continuous at the origin, then it is uniformly continuous in τ. In order to show this, let us consider the difference C τ +ε −=C τ EXtβ +⋅−−τ ;;;βββ⎡Xt ε Xt ⎤ XX() XX ()() ( )(⎣ )()⎦ Using the Schwarz inequality we obtain the inequality 2 −≤⎡ −⎤ CXX()τ +ε C XX ()τ 2C0C0C XX()⎣ XX () XX ()ε ⎦

Since CXX (τ) is continuous at the origin, C0CXX()−→ XX ()ε 0 as ε → 0 . Hence, the

difference CXX()τ +ε −C XX ()τ tends to zero with ε uniformly in τ.

4) As indicated by Eq. (3.20), CXX (τ) is nonnegative definite in τ.

5) Let CXX (τ) be continuous at the origin. It can be represented in the form 1 ∞ C τ = editω Φω (51) XX() ∫ XX () 2 −∞ If Φω( ) is absolutely continuous, we have

SddXX()ω= Φ XX () ω/ ω and 1 ∞ C τ = eSitω ωω d (52) XX() ∫ XX () 2 −∞ where S ()ω is real and nonnegative.

By virtue of Properties 2), 3), and 4) possessed by CXX (τ) this result is a direct consequence of the following important theorem due to Bochner.

Theorem. 2.44 : A function C()τ is continuous, real, and nonnegative definite if and only if, it has the representation (51), where ΦωXX () is real, non-decreasing and bounded. Proof : For the proof we refer to DOOB, J.L., (Stochastic Processes) ■

We see from Eq. (52) that the functions CXX ()τ and S XX ()ω form a Fourier transform pair. The relations 1 ∞ 1 ∞ C ()τ = eSitω ()ωω d SeC()ω = itω ()τ dτ (53) XX∫ XX XXπ ∫ XX 2 −∞ −∞ are called the Wiener-Khintchine formulas.

Equations (53) lead to the definition of the power spectral density function of a wide-sense stationary process whose correlation function is continuous.

44 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

Definition.2.45 : The power spectral density function of a wider-sense stationary s.p. X ()t ;β , tI∈ is defined by the first equation of (53), that is,

1 ∞ SeC()ω = itω ()τ dτ ■ XXπ ∫ XX −∞

We have already noted that the power spectral density function S XX ()ω is real and nonnegative. Furthermore, since CXX (τ) is an even function of τ, the Wiener-Khintchine formulas may be written as

2 ∞ ∞ SC()ωω= ()ττcos ( )dτ C ()τ = S (ωωω )cos (τ )d (54) XXπ ∫ XX XX ∫ 0 0

Hence, S XX ()ω is also an even function of ω. When τ = 0 , the second of Eqs. (54) reduces to ∞ Sdωω= C0 ∫ ()XX () 0 For a detailed discussion for stochastic processes we refer to ATHANASSOULIS, G.A., (Stochastic Modeling and Forecasting of Ship Systems) and PUGACHEV, V.S. & SINITSYN , (Stochastic Differential Systems).

2.2. CLASSIFICATION OF STOCHASTIC PROECESSES 45

2.2.3. Classification Based Upon Ergodic Properties

Ergodicity deals with the specific question of relating statistical or ensemble averages of a stationary stochastic process to time averages of its individual sample functions. The interchangeability of ensemble and time averages has considerable appeal in practice because, when statistical averages of a stochastic process need to be computed, what is generally available is not a representative collection of sample functions but rather certain pieces of sample functions or a long single observation of one sample function. One naturally asks if certain statistical averages of a stochastic process can be determined from appropriate time averages associated with a single sample function. Let x(t) be a sample function of a stationary s.p. X ()t ;β , tI∈=[]0, T. The time average of a t given function of x(t), gxt()(), denoted by gxt()() is defined by I

T /2 t 1 gxt()() = gxt()() dt (55) I ∫ T −T /2 if the limit exists. In order to define the ergodic property of a stationary stochastic process, the sample function x(t) is allowed to range over all possible sample functions and, for a fixed T, over all possible time translations; thus defines a random variable. Definition.2.46 : A s.p. X ()t ;β , tI∈ is said to be ergodic relative to G if, for every gG∈ ,

t t gXt⎡⎤();;ββ= Eβ () gXt ⎡⎤() ■ (56) ⎣⎦∞ ⎣⎦∞

It is, of course, entirely possible that, given a s.p. X ()t ;β , the ergodicity condition stated above is satisfied for certain functions of X ()t ;β and is not satisfied for others. In physical applications we are primarily interested in the following:

⎡⎤ 1. Ergodicity in the Mean: gXt⎣⎦();;ββ= Xt () The ergodicity condition becomes

t t Xt();;ββ= Eβ () Xt() (57) ∞ ∞

⎡ ⎤ 2. Ergodicity in the Correlation Function: gXt⎣ ();;ββ⎦ =⋅+ Xt ()( Xt τ ; β ). The ergodicity condition is

t t β t Xt()(;;ββ⋅+ Xt τ ) =EXt()()( ;; ββ ⋅+ Xtτ ) =RttXX () , +τ (58) ∞ ∞ ∞

For these ergodicity criteria, mathematical conditions can be derived for the purpose of verifying the ergodicity conditions. These conditions are direct consequence of the following

Theorem 2.47 [Ergodicity in the Mean] : Let X ()t ;β , tI∈ be an integrable, second order, t stochastic process and mT();;ββ= Xt (). We also assume that the correlation function X T

CttXX ()12, verify the condition

46 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

1 TT/2 /2 limCttdtdtXX ()12 , 1 2= 0 (59) T→∞ 2 ∫∫ T −−TT/2 /2 Then the ergodicity condition in the mean holds, i.e.

t t Xt();;ββ= Eβ () Xt() ∞ ∞ Proof : For the proof we refer to ATHANASSOULIS, G.A., (Stochastic Modeling and Forecasting of Ship Systems. Lecture Notes NTUA). ■

Theorem 2.48 [Statistical Estimation of mean value for a stationary s.p.] : Let X ()t ;β , tI∈ be a second order, m.s. continuous and stationary stochastic process. Then, for every ε > 0 and for every T > 0 , the following inequality holds,

T ⎛⎞t 1− Cdτ τ ∫ ⎜⎟XX () t 2 ⎝⎠T c Xt() −>≤⋅ m ε 0 (60) XX( T ) T ε 2

Concerning the Ergodicity criterion for the correlation function of a stochastic process the following theorem holds,

Theorem 2.48 [Ergodicity in the Correlation Function] : Let X ()t ;β , tI∈ be an t integrable, fourth order, stochastic process and Rˆ ()()()τ ,;TXtβββ=+⋅τ ;Xt ; . Let also XX T the correlation function CttYY ()12, τ of the stochastic process Yt()(,;τ βββ=+⋅Xt τ ; )()Xt ; , verify the condition

1 TT/2 /2 limCttYY ()12 , ;τ dt 1 dt 2= 0 (61) T→∞ 2 ∫∫ T −−TT/2 /2

Then the ergodicity condition in the Correlation Function holds, i.e.

t t Xt()(;;ββ⋅+ Xt τ ) =EXtβ ()()( ;; ββ ⋅+ Xtτ ) ∞ ∞

Proof : Direct application of Theorem 2.47 for the s.p. Yt()(,;τ βββ=+⋅Xt τ ; )()Xt ; ■ Example 2.49 : Consider the stochastic process

X (tA;cossin,ββ )=+ () tBttI () β ∈ where A(β) and B(β) are independent Gaussian random variables with means zero and variances σ 2 . The correlation function of X ()t ;β is

2 CttXX ()12,cos=−σ ( tt 1 2 )

It is easily shown that Eq. (57) is satisfied and thus the stochastic process is ergodic in the mean. ■ For a detailed discussion and more conditions for ergodicity criteria we refer to ASH, R.B., (Probability and Measure Theory) and ATHANASSOULIS, G.A., (Stochastic Modeling and Forecasting of Ship Systems. Lecture Notes NTUA). 2.2. CLASSIFICATION OF STOCHASTIC PROECESSES 47

2.2.4. Gaussian Stochastic Process

A discussion of stochastic processes is incomplete without giving considerations to Gaussian stochastic processes. Classification of stochastic processes based upon regularity and memory do not make specific references to the detailed probability distributions associated with a stochastic process. Gaussian stochastic processes are characterized by a distinct probabilistic law and they are extremely useful and most often encountered in practice.

Definition 2.50 : A stochastic process X ()t ;β , tI∈ is called a Gaussian stochastic process if for every finite set tt12,,… , tn ∈ I the random vector defined by the projection

Π ⎡⎤x()i…=∈{}xt ( ),, xt ( ) , x() i DI ( ) ()tt12, ,..., tN ⎣⎦ 1 n is a Gaussian random vector. ■

Thus, the probability density function of the random vector ()x()txt1 ,,… ()n , i.e. the n− joint probability density of the s.p. X ()t ;β is given by

fxxx,,,… = tt12, ,..., tN () 1 2 n

−nT/2 −1/2 ⎡ 1 ⎤ =−−−()2expπ CC⎢ ()()xm−1 xm⎥ (62) ⎣⎢ 2 ⎦⎥ where

T ⎡⎤⎡⎤ββ β m ==mmtt…… m t E() X t()ββ E() X t() E() X t() β (63) ⎣⎦⎢⎥12n ⎣⎦⎢⎥ 1 2 n and C = ⎡⎤μ is the nn× covariance matrix of X t ;β with ⎣⎦⎢⎥ij ()

β ⎡⎤⎡⎤ μββij=−−E Xt() i,, m t Xt() j m t (64) (⎣⎦⎢⎥ij⎣⎦⎢⎥)

Remark 2.51 : From the above equations it is clear that the probability structure of a Gaussian stochastic process is described by the moments of first and second order. ■

This is a typical characteristic of the Gaussian stochastic processes that makes them particularly useful for applications. Direct consequence of the above remark is that the higher order moments of every Gaussian characteristic function are expressed through the moments of first and second order. The theorem presented below, due to Isserlis 1918, give the necessary formula for the calculation of higher order moments.

Theorem 2.52 : Let the Gaussian stochastic process X ()t ;β , tI∈ . Then

I. The central moments of odd order, jk=+21, are identically zero, II. The central moments of even order, jk= 2 , are given by the formula,

jj12 jn CtttEXtmXtmjj12⋅ ⋅⋅⋅ jn , ,...,≡−β ;ββ ; − ... Xtm ; β −= XX... X() 1 2 n() () 1 t12( () 2 t) ( () n tn ) () = RttRtt,,… Rtt , ∑ XX() i12 i XX() i 34 i XX ( i 212kk− i )

48 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION where the summation is over all the possible partitions of the 2k indices ………… 1,1,,1;2,2,,2;,,, nn , n at pairs ()()(ii12,,,,, ii 34… i 2kk− 12 , i ). The number of terms is jj12 jn ()2!k =⋅⋅⋅135… ⋅() 2k − 1. ■ ()2!k k

For the Gaussian stochastic processes the following two important theorems hold,

Theorem 2.53 : Let the functions μ()i , I → and R(ii, ), I ×→I with the last one, satisfies the conditions

1) R()st,,,= Rts () for any ts, ∈ I 2) For any tt,,… , t∈ I the matrix ⎡Rt( , t)⎤ is non-negative definite. 12 n ⎣ mn⎦ nn×

Then, there always exists a Gaussian stochastic process X ()t ;β , tI∈ , with mean value function mt()= μ () t, tI∈ and covariance function R(ii, ), I ×→I , i.e. it satisfies the relations

Eβ ()Xt(),βμ= () t, tI∈ (65) Eβ ()Xt()(),,ββ Xs= Rts() ,, ts , ∈ I (66)

If the functions μ(i) and R(ii, ) are real then there is always a real stochastic process X ()t ;β , tI∈ that satisfies the above relations. In this case equation (66) becomes

EXtXsβ ()()(),,ββ= Rts() ,, ts , ∈ I ■ (67)

Theorem 2.54 : Let the stochastic process X ()t ;β , tI∈ with finite variance

EXtβ ()(),β 2 <∞, tI∈ ■ (68)

Then there is always exists a Gaussian stochastic process Yt( ;β) , tI∈ such that

mtYX()= m () t, tI∈ (69)

RYY()ts,,= R XX () ts, ts , ∈ I (70)

If the the stochastic process X ()t ;β , tI∈ is real then there is always a real stochastic process Yt();β , tI∈ that satisfies the above relations. ■

The above theorems are formulated and proved at the monograph of DOOB, J.L., (Stochastic Processes).

2.3. REFERENCES 49

2.3. References

ASH, R.B., 1972, Measure, Integration and Functional Analysis. Academic Press. ASH, R.B., 2000, Probability and Measure Theory. Academic Press. ATHANASSOULIS, G.A., 2002, Stochastic Modeling and Forecasting of Ship Systems. Lecture Notes NTUA. BROWN, R.G. & HWANG, P.Y.C., 1997, Introduction to Random Signals and Applied Kalman Filtering. John Wiley &Sons. CRAMER H. , LEADBETTER M. R.,1953, Stationary and Related Stochastic Process, Wiley NY. DOOB, J.L., 1953, Stochastic Processes. John Wiley &Sons. FRISTEDT, B. & GRAY. L., 1997, A Modern Approach to Probability Theory. Birkhäuser GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1996, Theory of Random Processes. Dover Publications. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1974, The Theory of Stochastic Processes I. Spinger. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1975, The Theory of Stochastic Processes II. Spinger. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1975, The Theory of Stochastic Processes III. Spinger. KLIMOV, G., 1986, Probability Theory and Mathematical Statistics. MIR Publishers. KOVALENKO, I.N. & KUZNETSOV, N.YU. & SHURENKOV, V.M., 1996, Models of Random Processes. CRC Press. KOLMOGOROV, A.N., 1956, Foundation of the Theory of Probability. LOÈVE, M., 1977, Probability Theory II, Springer. PERCIVAL, D.B. & WALDEN, A.T., 1993, Spectral Analysis for Physical Applications. Cambridge University Press. PROHOROV, YU.V. & ROZANOV, YU.A., 1969, Probability Theory. Springer. PUGACHEV, V.S. & SINITSYN , 1987, Stochastic Differential Systems. John Wiley &Sons. ROGERS, L.C.G. & WILLIAMS, D., 1993, Diffusions, Markov Processes and Martingales. Cambridge Mathematical Library. SINAI, YA.G., 1976, Introduction to Ergodic Theory. Princeton University Press. SNYDER, D.L., 1975, Random Point Processes. Wiley, New York. SOBCZYK, K., 1991, Stochastic Differential Equations, Kluwer Academic Publishers. SOIZE, C., 1993, Mathematical Methods in Signal Analysis. French, Masson, Paris. SOIZE, C., 1994, The Fokker-Planck Equation and its Explicit Steady State Solutions. World Scientific. SOONG, T.T., 1973, Random Differential Equations. Academic Press. SOONG, T.T. AND GRIGORIU, M., 1993, Random Vibration of Mechanical and Structural Systems. Prentice-Hall. SPILIOTIS, I., 2004, Stochastic Differential Equations. Simeon Press. YEH, J., 1995, Martingales and Stochastic Analysis. World Scientific.

50 CHAPTER 2 STOCHASTIC PROCESSES – DEFINITION AND CLASSIFICATION

Chapter 3

Probability Measures in Hilbert Spaces

As we have discussed at Section 1 of Chapter 2 the definition of the stochastic process as a measurable function which maps the space of elementary events Ω into the metric space [ lead us to the concept of a probability measure over an infinite dimensional space. In this chapter we will consider the case of probability measures defined on a Hilbert space. Every result presented above can be generalized to the case of a Banach space by replacing the inner product with the duality pair and making suitable changes and restrictions. In the first section we will give the definition of a stochastic process using probability measures. Moreover we will give some special examples of probability measures to get a clearer picture of the issue. Based on these examples we will introduce the finite dimensional relative measures, i.e. induced probability measures for finite dimensional subspaces of the Hilbert space. The next step will be to give the necessary and sufficient conditions for a family of probability measures, defined on every subspace of a Hilbert space, to define a probability measure. Section 2 deals with the special case of cylinder functions and connects their integral over an infinite dimensional Hilbert space, using induced probability measures for finite dimensional subspaces of the Hilbert space. In Section 3 we define the notion of the characteristic functional for probability measures over Hilbert spaces. Again, using finite dimensional relatives we make exact calculations of the characteristic functional. Finally, using nuclear operators we prove Minlos-Sazonov Theorem which is the generalization of Bochner’s Theorem for the infinite dimensional case and gives the necessary and sufficient conditions for a functional to be characteristic functional of some probability measure. In Section 4 we generalize the concepts of mean value and covariance of a random variable for probability measures defined over infinite dimensional spaces. As before, we can calculate this quantities using induced probability measures over finite subspaces. Section 5 and 6 present exact formulas for the calculation of classical moments and characteristic functions. The interesting part of this section is that we can calculate, using generalized functions the above quantities for the (time) derivatives of the stochastic process. Section 7 deals with special case of Gaussian measures and their properties. We prove that for this particular case of measure an exact formula for the characteristic functional. We also calculate the mean value and covariance operators and study their properties. In Section 8 a special case of characteristic functional is studied named the Tatarskii charactersitc functional. These functionals have the important property to represent a wide class of stochastic processes. Finally in section 9 we study the reduction of the Tatarskii characteristic functional to some very important characteristic functionals, such as the characteristic functionals due to Abel, Cauchy, and Feynman.

51 52 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

3.1. Probability Measures in Polish Spaces and their finite-dimensional relatives

3.1.1. Stochastic Processes as Random variables defined on Polish Spaces

In what follows we will consider the case of probability measures defined on a Polish space [ , i.e. a that has a countable dense subset. We also denote with [ ∗ the of linear continuous functionals on [ . We recall the definition of a stochastic process with values in [ .

Definition 1.1 : Let ()Ω, Uc()Ω , be a probability space, X :Ω→[ be a point function from the sample space Ω to the Polish space [ , and U[( ) be a Borel field over [ . The mapping X is said to be a stochastic process, iff is ()UU[()Ω , ( ) - measurable. ■

Any stochastic process induces a probability measure on the space it takes values.

Definition 1.2 : Let ()Ω, Uc()Ω , be a probability space, ()[U[, () a measurable space, and X :Ω→[ a stochastic process. On the Borel field U[() of the measurable space ()[U[, () we define the induced (by X :Ω→[ ) probability measure on U[() is defined by −1 cc[ ()BXB=∈( ()), for every B U[( ) (1)

The triplet ()[U[c,,()[ is called the induced probability space. ■

We shall now present some simple, specific examples of probability measures on Polish spaces. These examples are given by means of specific constructions permitting appropriate extensions of a finite-dimensional probability measure on the whole Polish space.

Example 1.3 : [The 1-D induced measure]: Let c be a probability measure in , and eˆ be a unit vector in [ . Let, also, ζ :[]eIRˆ → be a bijection between the span of eˆ , (denoted by [eˆ]) and the real axis , given by ζ (aeˆ) = a . Then, for any E ∈U[( ) we define ˆ cc[]eˆ ()EEe= ()ζ ()∩ [] (2) c[]eˆ is a probability measure in [ since, ˆ a) cc[]eˆ ()EEe=≥ ()ζ ()∩ [] 0 , E ∈U[() ˆˆ b) cc[]eˆ ()[[====()ζζ()∩ []ee c()()[] c() 1

c) for any countable collection of mutually disjoint sets EE12, ,..., inU[() ⎛⎞⎛⎞⎛⎞ ⎛⎞⎟ ⎛ ⎞ ⎜⎜EEeEeE⎟⎟===⎜ ζζ⎜ []ˆˆ⎟ ⎟ ⎜ [] ⎟() cc[]eˆ ⎜⎜∪∪∩∪∩nn⎟⎟⎜ ⎜ ⎟ ⎟ c ⎜() n ⎟∑ c[]eˆ n ⎝⎠nn⎜⎝⎠⎝⎠⎜ ⎝⎠⎟ ⎟ ⎝ n ⎠n 3.1. PROBABILITY MEASURES IN POLISH SPACES AND THEIR FINITE-DIMENSIONAL RELATIVES 53

[eˆ] ζ :[eˆ] →

E E2 E1 (),,Uc()

eˆ ,, ⊥2 ([U[() c[]eˆ )

eˆ⊥1 eˆ ⊥3 . ■

N Example 1.4 : [The N-D induced measure]: Let c N be a probability measure in , and LN be a linear subspace of [ , of dimension N . Let {ee12, , ..., eN } be a basis of LN , and N ζ :,,...,LeeeNN=→[ 12 ] IR be the bijection ζ (ae11+++= ae 2 2... aN eNN)( a 1 , a 2 , ..., a ). Then, for any E ∈U[( ) we define

cc⎡⎤()EEL= N ζ ()N . ■ (3) ⎣⎦⎢⎥ee12, ,..., eN ()∩

The set function c ⎡ ⎤ is a legitimate probability measure in [ obtained as an ⎣⎢ee12, ,..., eN ⎦⎥ extension of the N-D probability measure c N over the whole Polish space. In what follows we will examine the existence of a probability measure which is not an induced measure from a finite dimensional measure, but it has its own infinite dimensional structure. Before we procced we recall some basic results from probability theory.

3.1.2. Cylindric Sets

To begin our discussion we first need to give some essential definitions, concerning a special class of sets. Thus we have,

Definition 1.5 : A cylindric set on is defined to be a set QBn ()⊂ of the form [ l [ {}i i=1 nn QBQBxn ()==∈ () :,,,,,, lxlxlxB… ∈⊂∈ , B l l {}[ ()12 n U () {}i i=1 ∗ for every n =1, 2,… and l =()ll12,,… , ln with li ∈[ , in=1,… , . ■ ∗ The class of all cylindric sets for given li ∈[ , in=1,… , is obviousy a σ −algebra which we will write as . The union of all σ −algebras is an algebra. Indeed, if B and B U l n 1 2 {}i i=1 belong to and , then choosing lll= , we get, U l 1 U l 2 ()12

a) Bii ∈=U l ,1,2

54 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

∈ b) BB12∪ U l ∈ c) BB12∩ U l

d) BB12−∈U l

We denote this algebra of sets by U 0 . It is called the algebra of cylinder sets. Sets from U 0 are also considered as elementary for the construction of the measure on ()[ ,U[() as we shall see at the sequel.

3.1.3. Restriction of a probability measure over a Polish space, to finite-dimensional subspaces

We shall now study the restriction of a probability measure over a Polish space (assuming it exists), to finite-dimensional subspaces. Let (Ω, Uc(Ω), ) be a probability space, X :Ω→[ a stochastic process with values in a separable, real, Polish space [ , and U[() a Borel field over [ .

Definition 1.6 : Let ()[U[c,,()[ be the induced probability space form the s.p.

X :Ω→[ . Let ΠL be a projection of [ into the finite-dimensional subspace L . The probability measure defined on ()LL, U () as B ==∈∈∈ΠΠ−1 Bx: xBBL, (4) ccΠL () [[()LL[ ] c[({ [ ] }) U() is called the projection of the measure c[ on the subspace L . ■

Remark 1.7 : The fact that B is a measure follows form the fact that for a sequence of c ΠL () non-overlapping sets Bn ∈U ()L the sets

−−11⎡⎤ ΠΠLn⎢⎥∪∪B = Ln[]B ⎣⎦nn are also non-overlapping. ■

Remark 1.8 : Using a bijective mapping ζ : L → n where nL= dim[] we can define the above probability measure of the finite-dimensional subspace L on the borel field of n , as follows, −1 n ccn ()BBB=∈Π ()ζ (), U( ) ,ΠL L Hence, using the projection Π we define a r.v. into the probability space L, ()L , L ()UcΠL since the mapping defined is UU()Ω , n -measurable. We denote this r.v. as Ξω( ).■ ()() ΠL

3.1. PROBABILITY MEASURES IN POLISH SPACES AND THEIR FINITE-DIMENSIONAL RELATIVES 55

([U[,,( ) c[ )

,,Uc() ,P ( []e1 )

eˆ2 −1 ζ : → [eˆ1 ]

eˆ1

The measures c Π , being projections of the same measure, are compatible in a certain sense, Ln for different n. This compatibility condition is the generalization of the consistency condition of Kolmogorov Theorem (Chapter 2/ Theorem 1.7).

Compatibility Condition 1.9 : Let the probability measure c[ defined on the measurable space , and the family of measures defined for all finite-dimensional ()[U[() {c ΠL } subspaces L of [ . Then for all LL12⊂ and B ∈U (L1 ) we have,

B = Π −1 BL (5) ccΠΠ() ()L1 []∩ 2 LL12

Proof : Let LL⊂ and B ∈ L . Then the set Π −1 B can also be written as Π −1 B , 12 U ()1 L1 [ ] L2 [ 2 ] where B22∈U ()L is defined by the equality

B =∈xL:Π x ∈ B 22{ L1 [] } Since ΠΠ−−11B = B , we have LL12[ ] [ 2 ]

B ==ΠΠ−−11BBB = ccΠΠ() [[()LL12[] c( []22) c() L1 L2

Since B =∈xL:Π x ∈ B, we have prove the condition (5). ■ 22{ L1 [] }

3.1.4. Conditions for the existence of probability measures over a Polish space

We will now study the inverse problem, that is to investigate the conditions under those, a family of probability measures c Π , for a sequence Ln of linear subspaces of [ , for Ln which LLnn⊂ +1 and ∪ Ln is dense in [ , can define a probability measure on [ . This n family of probability measures is redundance, so we cannot associate it with a unique probability measure of [ , unless it satisfies the Compatibility Condition 1.9. Using the

56 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

arguments of Example 1.4, one can determine on L , by knowing . Hence c[ U ( ) c ΠL knowing c Π , for a sequence Ln of linear subspaces of [ , we can define c on Ln [

∪U ()Ln and since the σ-closure of this algebra coincides with U[( ) we can define c[ n by the same token. Unfortunately, there are extra requirements for a family of probability measures, satisfying Compatibility Condition 1.9, to correspond to a probability measure on ()[U[, (). The next theorem gives a necessary and sufficient condition for a weak distribution to correspond to some measure.

Theorem 1.10 : Let Sρ be a sphere of radius ρρ::Sxxρ =≤{}. A family of probability measures , satisfying Compatibility Condition 1.9, will be generated by some measure {c ΠL } c[ onto the measurable space ()[U[, ( ) iff for every ε > 0 there exists an η > 0 such that for all L

SL≥−1 ε when ρη> (6) c ΠL ()ρ ∩

Proof : Necessity. If is generated by the measure , then choosing η > 0 such that {c ΠL } c[ c Sη >−1 ε (this is possible since limccSη = ([ ) =1), we obtain [ () η→+∞ [[( )

SL= Πε−1 SL≥≥>− S S1 . ccΠL ()ρρρη∩∩[[[()L () cc( ) ( )

The proof of sufficiency is more difficult. We define on the algebra UU0 = ∪ ()Ln a finitely n additive function c[ by means of

AAA=⊂, . cc[ ()ΠL () U0

To convince ourselves that c[ can be extended to a measure defined on ()[U, [ , it is sufficient to show that c[ is continuous on U 0 , i.e., that for an arbitrary sequence of sets

An ∈U 0 for which AAnn⊃ +1 and ∩ An =∅, n

limc ()An = 0 (7) n→∞ [

Let An be a cylinder set with base in Ln and LLnn⊂ +1 . Let Bnn⊂ L be the base of An . We remark that it is sufficient to prove (7) merely for sets with closed bases. Indeed, choosing closed sets CB⊂ such that BC− =−< BC ε and then taking nn ccΠLn()nn Lnn( ) n

n DxxCL=∈:Π , nLmn∩∩{}m m=1 we obtain closed sets for which

nn BC−≤ BC −≤ε ccLnnnm()∑∑ L ( mm ) m mm==11

3.1. PROBABILITY MEASURES IN POLISH SPACES AND THEIR FINITE-DIMENSIONAL RELATIVES 57

Hence, if AD′ = Π −1 (), then nLnn

n AA′′⊂ , AA′ = and AA≤+′ ε nn+1 ∩∩nncc[[()nnm( ) ∑ nn m=1 ∞ ′ If for cylinder sets with closed bases limc An = 0, then since ε n can be taken as n→∞ [ () ∑ n=1 suitably small, (7) is also fulfilled. Hence, it will be assumed that the Bn are closed sets. Then w. the An will be weakly closed sets (if xk ⎯⎯→ x and xkn∈ A , then x ∈ An ). For all ρ the sets

Sρ is also weakly closed and weakly compact. Since,

⎡⎤∞ SA⎢⎥=∅ ρ ∩∩⎢⎥n ⎣⎦n=1

∞ ∪∩SAρ n =∅ and since the sets SAρ ∩ n are weakly closed and weakly compact and n=1

SASAρρ∩∩nn⊃ +1 , for some n , SAρ ∩ n =∅. This means that, AALLS=≤− ≤ε cc[ ()nLnLnLnnnn () c( ) c( ∩ ρ ) provided that ρη> . From the arbitrariness of ε > 0 there follows (7). ■

To understand more precisely the role of condition (6) let us note that, for an arbitrary family of measures, is equivalent with the property of “uniformly tight” described by the following

Definition 1.11 : A family {c } of probability measures on [ is uniformly tight if, for every ε > 0 there exists a compact subset ^ of [ such that c ()^ ′ < ε for all cc∈{ } . ■

Concerning the property of uniformly tightness the following important theorem holds,

Theorem 1.12 [Prohorov] : A family {c } of probability measures on a Polish space [ is relatively sequentially compact if and only if is uniformly tight. ■

Hence condition (6) assures that for every sequence c n (n =1, 2,…) of members of {c }, there exists a convergent subsequence. Now consider a family of probability measures

c Π , for a sequence Ln of linear subspaces of [ , for which LLnn⊂ +1 and Ln is { Ln} ∪ n dense in [ . Then, to assure the uniqness of the limit of every convergent subsequence we need Compatibility Condition 1.9. Indeed, if compatibility condition 1.9 holds and for two subsequences there are two different limits, it can be easily seen that we are leading to an absurd. Thus, the property of “uniformly tight” for a family of probability measures, assures the existence of a convergent subsequence, while, the compatibility condition assures the uniqness of the limit for every convergent subsequence. Finally the existence of a probability measure is assured by the following

58 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

Preposition 1.13 : Let c n ()n =1, 2,… be a relatively sequentially compact sequence of probability measures on a Polish space such that every convergent subsequence has the same limiting probability measure c . Then ccn → as n →∞. ■

Example 1.14 : Let [ be the Cab([ , ]) space of continuous functions on the interval [ab, ] and U (C) the σ −field generated by the open subsets of Cab()[], . Let Π :,Cab()[]→ IRN be the projection defined by ()tt12, ,..., tN Π ⎡⎤x()i…=∈{}xt ( ),, xt ( ) , x() i C() [ ab , ] ()tt12, ,..., tN ⎣⎦ 1 n For the moment we assume that c exists and it’s a probability measure over Cab()[], . We thus can define the sequence of finite probability measures

⎡ ⎤ cc()EE==()Π()tt, ,..., t Cab()[], ()tt12, ,..., tN ∩ 12 N ⎣⎢ ⎦⎥

=∈...f ()xx , ,..., x dxdx ... , E () C ∫∫tt12, ,..., tN 1 2 n 1 N U AA ttN 1

Now, may be thought of as the limit of the function ...f ()xx , ,..., x dxdx ... , c ∫ ∫ tt12, ,..., tN 1 2 n 1 N AA ttN 1 which has discrete arguments. The above limit exists and it is a probability measure if and only if the family of distributions f (xx, ,..., x) fulfils the conditions of Kolmogorov tt12, ,..., tN 1 2 n theorem (Chapter 2/ Theorem 1.7), which are special case of the conditions presented above, for general measures. ■

Remark 1.15 : We note that for the definition of the probability measure c[ on the measurable space , it is sufficient to know the family of measures ()[U[() {c Lz } defined for all one-dimensional subspaces L of [ spanned by a vector z ∈[ . Thus, the measure c[ is defined by its one-dimensional distributions (i.e. by its projections onto one- dimensional subspaces). The proof of the above basically relies to the characteristic functional approach that will be studied in the next section. Using this remark we could define characteristic functionals by knowing their one-dimensional distributions. However this cannot be done in general, since the resulted measure must satisfy the definition of a probability measure, a condition which is not, in general, fulfilled.

3.2. CYLINDER FUNCTIONS AND THEIR INTEGRAL 59

3.2. Cylinder functions and their Integral

It is interesting to note that some classes of functions can also be integrated with respect to some projection of the probability measure, defined on the Hilbert space. To these functions, for example, are related the cylinder functions defined below.

Definition 2.1 : A function Z[ : →V , where V 1 is a linear space, is called a cylinder function if there is a finite dimensional subspace L ⊂[ and a finite dimensional U (L) - measurable function gLL : → V, such that

Z[()xg=∈LL()Π [] x, x (1)

where ΠL is the projection of [ on L . ■

In a Hilbert space, a finite dimensional projection ΠL [i] can be represented as

{}ly1,,,,… lyn where li are n linearly independent elements of the Hilbert space. With this form of projection it’s possible to integrate cylinder functions of the form gly()1,,,,… lyn which will play an important role for the sequel. Let ()Ω, Uc()Ω , be a probability space, X :Ω→[ a stochastic process from the space of elementary events Ω to the separable, real, Hilbert space [ and U[() a Borel field over [ . Let also, c[ be the induced probability measure. For every n , and for every n vector element l =ll{}12,,… , ln ∈[ we consider the projection Πl :[ → L ,

ll12 ln Πl []xlx=+12,, lx ++∈… lxn , x[ (2) ll12 ln and the injection ζ : L → n defined by,

⎛⎞n ζ ⎜ al⎟= {} a,,, a… a (3) ⎜∑ ii⎟ 12 n ⎝⎠i=1

Based on the above discussion we have the restriction of the probability measure c[ onto n ,

n cc[l ()Ex=∈[ ( :{ lylyEE1,,,,… n } ∈) ∈ U( ) (4) or

n FFaall()aa==∈≤=∈ (1,,……nii ) c[[ () x: lxain , , 1,, (5)

We then have the following

Definition 2.2 : Let g : n → be an arbitrary integrable, measurable function, defined in n n , with values in . For every n , and for every vector l =ll{}12,,… , ln ∈[ with li ≠ 0 for

1 In our applications V is usually or .

60 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

every in=1,… , , we define the infinite dimensional integral of the function g : n → by the following relation

g l,,,, x……… l x dx= g t ,, t dt ,, dt (6) ∫∫()111nnncc[ () ( )l ( ) [ n where, cl is defined by eq. (4).

Remark 2.3 : The above definition physically connects the integral over an infinite dimensional space with a finite dimensional subspace. To understand more precisely the above connection suppose that in some way we could define the infinte dimensional integral in the l.h.s. of equation (6). Then we would have Πl []x = {}lx1,,,,… lxn . Making the set valued transformation

−1 n x =→Πl []t : [ we have,

g P x dx= g t,, t…… , t P−1 dt dt dt ∫∫()ll[]cc[[() (12nn )() [ 1 2 ] [ n

Using the definition of eq. (4) we see that

−1 n cc[[[()Πl [ExlxlxEEE] =∈():{ 1,,,,… n } ∈= cUl () ∈() . Thus, the reduction of the infinite dimensional integral is due to the fact that the integrable function g : n → depends from a finite dimensional subspace of [ . ■

Remark 2.4 : As was illustrated above, one of the most important applications of the projection of a measure is the calculation of integrals of cylinder functions. These finite- dimensional measures are adapted to the form of the integrable cylinder function, using a suitable projection. The above current of ‘information’ can be summarized to the diagram below,

(Ω, Uc(Ω), ) ⎯⎯⎯⎯⎯ ⎯ Ξ i) ( ( i X ) → ←⎯ ⎯ ⎯⎯ [U[c,,() ⎯⎯⎯⎯⎯{ ll1 ,,i… ,n , i}→ nn,, U c ()[ ( ( ) l ) ■

Remark 2.5 : For n =1 the above definition reduce to the equation

glx, dx=∈ gt dt l (7) ∫∫()cc[() ()()l [ where

cc[l ()axlxa=∈{}; , ≤ for every la ∈ [ , ∈ (8)

The above equation will be used extensively in what follows. ■ 3.3. CHARACTERISTIC FUNCTIONAL OF THE PROBABILITY MEASURE 61

3.3. Characteristic Functional of the Probability Measure

In this section we will generalize the concept of the characteristic function for probability measures on real, separable, Hilbert spaces. First we recall the definition of a positive definite functional.

Definition 3.1 : A function φ from [ into (or ) is called a functional of positive type if for any xx12, , ..., xn ∈=[ , n 1,2, ...and any numbers cc12, ,..., cn ∈ we have:

n ∑ φ()xxccjkjk−⋅⋅≥0 ■ (1) jk,1=

We are, now, going to define the characteristic functional for a probability measure. Let [ be a real, separable Hilbert space and U[() the σ −field generated by the open subsets of [ . Thus, we have the following

Definition 3.2 : The characteristic functional Y :[ → of a probability measure c[ on [ , is defined by

Yc[()xEe==β ()ix,, y∫ e ix y () dyx ∈ ■ (2) [ Remark 3.3 : Since c[()<∞, Y always exists. Moreover,

Yc[[()xx≤=∀∈()1 ■ (3)

Remark 3.4 : Using the distribution of the measurable function x,i and the remark 2.5. we have that ∞ xe==ix, y dyedtxit ∈ (4) Ycc[() ∫∫()x () [ −∞ Hence, this is how the probability measure can be defined, by using the family of measures defined for all one-dimensional subspaces L of spanned by a vector x ∈ . {c Lx } [ [ (See Remark 1.12). ■

Concerning the smoothness of Y we have the following theorem.

Theorem. 3.5 : Let c be a probability measure in [ and let Y be the characteristic functional of c . Then Y is uniformly continuous. ε Proof : Let ε > 0 be given. Then there exists r > 0 such that c ()Sr >−1 , where 4

Sxxrr =≤{}: . Then

YY()x −≤ ()yee∫ ixz,, − iyz c() dz = [

62 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

=−∫∫eeixz,, iyzcc() dzee +− ixz ,, iyz () dz ≤ S ′ r Sr ε ≤−xz,, yzc () dz +⋅≤ 2 ∫ 4 Sr ε ≤−+rx y . 2 ε Choose δ > 0 such that δ < . Then, xy−<δ implies YY()xy−< () ε . Hence, is 2r uniformly continuous. ■

If [ is an infinite dimensional Hilbert space we can show something more for Y . For the sake of convenience we introduce the following

Definition 3.6 : An operator is called a of [ if ∞

1. it has finite trace, i.e. , for some orthonormal basis {}ei , ∑ Seii, e <∞, i=1 2. it is self – adjoint ,

3. it is positive – definite. S denotes the nuclear operators of [ . ■

This theorem gives necessary and sufficient conditions for a functional Y be the characteristic functional of some measure on ()[U[, (). It is thus a generalization of the Bochner’s theorem on positive definite functions in Hilbert space.

Theorem. 3.7 [Minlos-Sazonov] : A functional Y in [ is the characteristic functional of some probability measure in U[() if and only if a) Y ()01= and Y is a functional of positive type,

b) For every ε > 0, exists Cε ∈ S such that

1−≤+ ReY[()xCxxε ,ε for all x ∈. (5)

Remark. 3.8 : b) implies, in particular, that when Cxxε , is small, then 1Re− Y ()x is small. Note that Cxxε , can be small even if x is big. Note also that {}Cxxε , < a is an 11 22 ellipsoid with semi-axes ()aa/λλ12 ,() / , ... , whereλn ’s are the eigenvalues of Cε .

Proof : Necessity. Let Yc()x = ∫ edyix, y () and ε > 0 be given. Choose 0 <<∞r such [ that S >−1 ε where S is the ball x ≤ r . Then c ()r 2 r {}

Ycc()x =+∫∫edyedyix,, y() ix y () SSrr′ 3.3. CHARACTERISTIC FUNCTIONAL OF THE PROBABILITY MEASURE 63

Note that

edySix, y ≤<′ ε ∫ cc() ()r 2 Sr′

Hence it is sufficient to show that

ix, y ε 1Re−≤+edyCxxc () ε , ∫ 2 Sr for some operator Cε ∈ S . But

1Re−=−+≤edyxydySix, y ⎡⎤ 1cos, ′ ∫∫ccc() ⎣⎦() ()r SSrr ε ≤−⎡⎤1cos,xyc () dy + ∫ ⎣⎦2 Sr 1 Recall that 1cos−≤θθ2 for all real θ . Hence 2

1 2 ⎡⎤1cos,−≤x ydyxydycc() , () ∫∫⎣⎦2 SSrr

The same argument in the proof of Theorem 4.10 below shows that there exists Cε ∈ S such that

1 Cxxε ,,,= xz yzc () dz 2 ∫ Sr

The desired conclusions follow immediately.

Remark : Observe that if the covariance operator Cc of c exists, then we have

1Re−≤()xCxxx1 , for all ∈ Y[2 c

But Cc , in general, is not an S - operator. Of course, if Cc ∈ S , then b) is trivially satisfied 1 by taking CC= for all ε > 0. ε 2 c

Remark : Observe also that Cε is the covariance operator of the following Borel measure c in [ ,

1 cc()EESE=∈()r , U()[ 2 ∩

Clearly c satisfies the hypothesis of Theorem 4.10 and thus its covariance operator Cε is an

S - operator. This is another proof for Cε ∈ S .

Sufficiency. (We will need to use Bochner’s Theorem for the finite dimensional space n )

Step 1. We first derived some properties implied by a).

64 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

a1) Y ()x ≤1 , and YY(x)=−( x) for all x ∈[ , a2) YY()x −≤−− ()yxy21 Y() for all xy, ∈[ , a3) 121Re−≤−YY()x ()x for all x ∈[ .

a1) : Take n = 2 , x1 = 0 and x2 = x . Part a) implies that the following matrix is positive definite in the sense of linear algebra,

⎛⎞ ⎜ 1 Y ()x ⎟ ⎜ ⎟ ⎝⎠⎜Y ()−x 1 ⎟

Hence YY(x)=−( x) . Moreover the matrix determinant is non-negative, i.e. 10−YY()xx ()≥ . Hence Y ()x ≤1.

a2) : Take n = 3, x1 = 0 , x2 = x and x3 = y . Part a) implies that the following matrix

⎛⎞ ⎛⎞1 YY()xy () ⎜ 1 YY()xy () ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ ⎟ ⎜YYYY()−−=−x 11 (yx )⎟ ⎜ () x() xy⎟ ⎜ ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ ⎟ ⎝⎠⎜YY()−−yxy ( ) 1 ⎝⎠⎜YY()yxy()− 1 ⎟ is positive definite. Hence the matrix determinant D ≥ 0 . But

Dxyxyxyxyy=+1 YYY() () () − + YYY() () ( − ) − Y ()2 −

−YY()xxy22−−= ( ) 22 2 =+12Re⎡⎤()xyxy () () − −() x +() y −() xy − = ⎣⎦⎢⎥YYY() Y Y Y 22 =+12Re⎡⎤()xyxyx () () − −() − () y −() xy − − ⎣⎦⎢⎥YYY Y Y Y

−=2Re⎡⎤()xy () ⎣⎦⎢⎥YY 22 =+12Re⎡⎤YY()x ()yxy Y() − − 1 − Y() x − Y () y − Y() xy − ⎣⎦⎢⎥()

Note that 2 11121−−=−−+−≤−−YYYY()x yxyxyxy()()()() ()

and

2Re⎡⎤YY()xy () Y() xy−−≤ 1 2 Y() x Y() y Y() xy −−≤ 1 ⎣⎦⎢⎥() ≤−21 Y ()x −y i.e. YY()x −≤−− ()yxy21 Y()

a3) : Note that if z ≤1 then 3.3. CHARACTERISTIC FUNCTIONAL OF THE PROBABILITY MEASURE 65

()1−=−−=−++≤−zzzzzzz2 () 1() 1 1()2 12Re1 +=− 21Re() z From a1) we know that Y ()x ≤1, therefore,

2 121Re−≤−YY()x ()()x

Hence,

121Re−≤−YY()x ()x . Remark : Property a2) says that if Y ()x is continuous at the origin (w.r.t. whatever topology) then Y ()x is continuous in the whole [ (w.r.t. the same topology). Property a3) says that the continuity of ReY ()x impies the continuity of Y ()x . Compare b) of this Theorem.

Step 2. Let {}e be a fixed orthonormal basis of . For each n ≥ 1, define ()x in n [ Y ee1 ,,… n n by ()(aa,,……=++ aeae ) YYee1 ,,… n 1 n 11 nn

Note that ReY is continuous at the origin by b). Hence Y is continuous in [ by a2) and a3). Hence is positive definite and (0,… ,0)= 1. Bochner’s Theorem for n Y ee1 ,,… n Y ee1 ,,… n gives us a family of probability measures {}c n such that

()aa,,… = eiay⋅ () dy , a =()aa,,… Ycee1 ,,… n 1 n∫ n 1 n n

It is easy to see that {}c n is a consistent family. Hence Kolmogorov’s Theorem implies the existence of a probability space (Ω, cΩ ) and a sequence of random variables {}X n such that

−1 ccnn==Ω ()XX12,,,…… X , n 1,2,

Therefore,

ae+++= ae… ae eiay⋅ dy = , a = aa,,… Yc()()11 2 2 nn∫ n ()()1 n n iaX( +++ aX… aX ) = ∫ ed11 2 2 nn c . Ω

∞ 2 Step 3. Suppose we can show that ∑ X n <∞ almost surely (this will be shown in Step 4) n=1 then we are done. To see this define,

∞ XXe()ωωω=∈Ω∑ nn () , n=1

− 1 Then X is measurable from Ω into [ . Define cc= Ω X . c is a probability Borel measure of [ . Let Πn be the orthogonal projection of [ into the span of ee12,,… , en ; i.e.

Πnnnx =+xe,,11 e xe 2 e 2 ++… xe , e , x ∈[ .

66 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

Then

n ΠnkkX = ∑ Xe k=1

By Step 2 we have,

Π Π xed= ix, n X . Yc()n ∫ Ω

Now letting n →∞ and observing Πn x → x in [ as n →∞, we have Y ()Πn x by the continuity of Y . Apply Lebesgue’s Dominated Convergence Theorem,

Π ∫∫edix, n X cc→ edixX, as n →∞ ΩΩ

Hence we have,

Yc()xed=∈∫∫ixX,, = e ix yc[() dyx, . Ω [

∞ 2 Step 4. To show that ∑ X n <∞ almost surely. n=1 First note that

n 11 −22 + ++…… 2 − 22 + ++ 2 iay()+++ ay… ay ⎛⎞1 ()yy12 ynn() aa 12 a e11 2 2 nn ⎜ ⎟ e22 dy dy… dy= e ∫ ⎝⎠⎜ ⎟ 12 n n 2π

Therefore

1 22 2 −+++(XXkk++12… X kn +) ∫ ed2 c = Ω ⎡⎤n ⎛⎞⎛⎞nn ⎛⎞11⎟⎟2 =−⎢⎥⎜ ⎟ exp⎜⎜i X Y⎟⎟ exp Y dy dy… dy d = ∫∫⎢⎥⎜ ⎟ ⎜⎜∑∑kjj+ ⎟⎟ j12 n c n ⎝⎠2π ⎜⎜2 Ω ⎣⎦⎢⎥ ⎝⎠⎝⎠jj==11 ⎡⎤⎛⎞nnn ⎛⎞ ⎟⎟⎛⎞112 = ⎢⎥⎟exp⎜⎜iXYd⎟⎟⎜ ⎟ exp −= Ydydydy… ∫∫⎢⎥⎜⎜∑∑kjj+ ⎟⎟c ⎜ ⎟ j12 n n ⎜⎜⎝⎠2π 2 ⎣⎦⎢⎥Ω ⎝⎠jj==11 ⎝⎠ n ⎛⎞11⎛⎞n ⎜ ⎟ ⎜ 2 ⎟ = Y ()y11ekk+++++ y 2 e 2…… y nkn e +⎜ ⎟ exp⎜ − y j⎟ dy 12 dy dy n = ∫ ⎝⎠⎜ ⎟ ⎜ 2 ∑ ⎟ n 2π ⎝⎠j=1 n ⎛⎞11⎛⎞n ⎡⎤⎜ ⎟ ⎜ 2 ⎟ = Re Y ()y11ekk+++++ y 2 e 2…… y nkn e +⎜ ⎟ exp⎜ − y j⎟ dy 12 dy dy n , ∫ ⎣⎦⎢⎥⎜⎝⎠⎟ ⎜ 2 ∑ ⎟ n 2π ⎝⎠j=1 since the left hand side is real.

Now let ε > 0 be given, we use assumption b) to get Cε ∈ S such that

1−≤+ ReY[()xCxxε ,ε for all x ∈.

Therefore,

3.3. CHARACTERISTIC FUNCTIONAL OF THE PROBABILITY MEASURE 67

1 22 2 −+++(XXkk++12… X kn +) 1−∫ ed2 c ≤ Ω ≤ 1Re−+++⎡⎤ye ye… ye p dy , ∫ ()⎣⎦⎢⎥Y ()11kk++ 2 2 nkn + 1() n where ⎛⎞n ⎛⎞ ⎜ 11⎟ ⎜ 22 2⎟ p11212()dy =−+++⎜ ⎟ exp⎜ () y y…… ynn⎟ dy dy dy ⎝⎠⎜ 2π ⎟ ⎜⎝⎠2 ⎟ is a probability measure in n . Then we have,

1Re−+++≤⎡⎤ye ye… ye p dy ∫ ()⎣⎦⎢⎥Y ()11kk++ 2 2 nkn + 1() n ≤+++++++=⎡⎤Cyeyeyeyeyeyepd……, ε y ∫ ⎣⎦⎢⎥ε ()11kk++ 22 nknkk +++ 11 22 nkn + 1() n nn =+ε yy Ce, e p dy = ∫ ∑∑ ijε kikj++ 1 () n ij==11 nn =+ε Ce, e yy p dy = ∑∑ ε ki++ k j∫ i j 1 () ij==11 n n =+ε ∑ Ceε kj++, e kj. j=1

Let n →∞, apply Lebesgue’s Dominated Convergence Theorem to the left hand side and remember that Cε ∈ S . Hence,

1 ∞ − X 2 ∞ 2∑ kj+ 1,−≤edj=1 Cee ∫ c ≤+ε ∑ ε mm Ω mk=+1

≤ 2ε whenever kk≥ 0 say.

Hence,

1 ∞ − X 2 2∑ kj+ edj=1 ≥−12ε , kk≥ . ∫ c 0 Ω

Finally,

1 ∞ ∞ − X 2 ⎪⎪⎧⎫2∑ kj+ ⎪⎪⎨⎬Xed2 <∞ ≥j=1 ≥12 − ε cc⎪⎪∑ n ∫ ⎩⎭⎪⎪n=1 Ω

Thus,

22 c {}XX1 ++……n +<∞≥−12ε , for any ε > 0.

Of course we must have

22 c {}XX1 ++……n +<∞=1, which is what we wanted to prove. ■

68 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

Finally we present some basic properties concerning operations with characteristic functionals

Theorem 3.9 : a) If Y 1 ()x and Y 2 ()x are characteristic functionals and pp12+=1, where

p12, p are non-negative numbers, then p11YY()xp+ 2 2 () x is also a characteristic functional. b) If Y 1 ()x and Y 2 ()x are characteristic functionals, then YY12()x ⋅ ()x is also a characteristic functional. c) If Y s (x) is a family of characteristic functionals depending upon a random parameter s , and Fs() is the distribution of s , then x dF s ∫ Y s () () S is also a characteristic functional. d) Let Y X ()x be a characteristic functional of the process X ()t ;β and Y Y ()x a characteristic functional of another process Yt( ;β) being a linear transformation of X ()t ;β . d1) If Yt();;ββ=+ Xt () at (), where at( ) a non-random function, then ⎧⎫ ⎪⎪ ()x =⋅ ()xiatxtdtexp⎨⎬ ()() YYYX⎪⎪∫ ⎩⎭⎪⎪T

d2) If AL is an of the form Yt;;,;ββ== AXt GstXs β ds ()L ()∫ ()() T where Gst(), is a suitable non-random function, then ⎛⎞ ∗ ⎜ ⎟ YYYXLX()x ==()Ax Y⎜ Gtsxsds()(), ⎟ ⎜∫ ⎟ ⎝⎠T ∗ where AL is a conjugate to AL .

Proof : Properties a) and c) follows from the corresponding properties of the probability measures and the continuity properties of the infinite dimensional Fourier transform. Property b) is a direct consequence of the linearity property of the Stochastic Processes space, since YY12(x)⋅ (x) is the characteristic functional of the sum X12()tXt;;ββ+ () with characteristic functional YY12()x , ()x respectively. Finally, properties d) can be proven immediately using the definition of the characteristic functional. ■

Remark 3.10 : The last group of properties are very essential tools for the combination of different characteristic functionals, to create new ones, more efficient for solving Functional Differential Equations. As we shall see at Chapter 5 the simplest application of those properties is the convex superposition of Gaussian Functionals to make a representation with local characteristics (kernel characteristic functionals). ■

3.4. MEAN AND COVARIANCE OPERATOR FOR PROBABILITY MEASURES 69

3.4. Mean and Covariance Operator for Probability Measures

Now we are going to generalize the concept of the covariance functions for a probability measure defined on a real, separable, Hilbert space [ .

Definition 4.1 : Let c be a probability measure in [ . The mean value of c is an element mc ∈[ such that

mx,,=∈ zx dzx ■ (1) c ∫ c[() [

Remark 4.2 : In general m does not exist. However, if x dx then m exists and c ∫ c () c [ mxdx≤ ■ (2) c ∫ c () [

Remark 4.3 : Using the distribution of the measurable function x,i

ccx ()ayxya=∈{}[ ;, ≤

We have that

∞ mx,,==∈ zx dz t dtx ■ (3) c ∫∫cc[()x () [ −∞ Analogously we define the covariance operator.

Definition 4.4 : Let be a probability measure in . The covariance operator C of c [ c c is defined by

xy,,,,=∈ x z y z dz xy ■ (4) Cc ∫ c[() [

Remark 4.5 : The operator Cc may not exist. If Cc exists, it is positive definite and self – adjoint. ■

Remark 4.6 : Using the distribution of the measurable function x,i

cc[x ()ayxya=∈{};, ≤ (5)

We have that

∞ xx,,== x z2 dz t2 dt xy , ∈ ■ (6) Cc ∫∫cc()x () [ [ −∞

Example 4.7 : The covariance operator of δx is Sxδ = xxx, 00 since 0 x0

Cxyδ ,,,,,== xx00 yx xxxy 00 ■ (7) x0

Example 4.8 : The covariance operator of c e (example 1.4.) is

70 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

⎛⎞ ⎜ 2 ⎟ C x =⎜ tdtxeec ()⎟ , ˆˆ (8) c e ⎜∫ ⎟ ⎝⎠IR since,

Cxy,,,=== xzyz() dz xzyz ,,() dz c e ∫∫ccee [ []eˆ ˆˆ ˆ2 ˆˆ2 ==x,,eye∫∫ ze ,cce () dzxeyet ,, () dt ■ []eIRˆ

⎛⎞⎟ Remark 4.9 : Suppose that ⎜ tdt2 ()⎟=1. Then CC= even thought and δ are ⎜ c ⎟ c e δ c e x0 ⎜∫ ⎟ x0 ⎝⎠IR two different measures. Hence c e is not uniquely determined by its covariance operator. However, we will see later on that a Gaussian measure of mean zero is completely determined by its covariant operator. ■

The following theorem gives us all the necessary condition for an operator to be the covariance operator of a probability measure.

Theorem 4.10 : Let c be a probability measure in [ and Cc the covariance operator of . Then xdx2 <∞ if and only if ∈ S . In fact, trace ⎡ ⎤ = x 2 dx . c ∫ c () Cc ⎣⎦Cc ∫ c () [ [

Proof : Sufficiency. Let {}en be an orthogonal basis of [ . By Monotone Convergence Theorem we have,

2222⎡⎤ x cc()dx=++lim xe ,12 xe ,… xe , n () dx ∫∫n→∞ ⎣⎦⎢⎥ [[

But

2 x,,edxee= ∫ j c () Cc jj [

Hence,

n ∞ 2 xdxc ()===lim CeeCeejj , jj , ∫ n→∞ ∑∑cc [ jj==11

=<∞trace[]Cc

Necessity. First of all, we have to show the existence of the covariance operator. Clearly,

x,,zyzxyz⋅≤⋅⋅2

Hence, 3.4. MEAN AND COVARIANCE OPERATOR FOR PROBABILITY MEASURES 71

∫∫x,,zyz⋅≤⋅⋅cc() dzxy z2 () dz [[

Therefore the bilinear form ∫ x,,zyz⋅ c () dz is continuous. Hence there exists a linear [ operator Cc in [ such that,

x,,,yxzyzdz=⋅ Cc ∫ c () [

∈ Obviously, Cc is self-adjoint and positive definite. To show that Cc S , it is sufficient to ∞ show that if {}en is an orthonormal basis of [ then the series ∑ Ceec j , j is convergent. j=1 But

∞∞ 2 Cee,,== xe dx ∑∑c jj∫ jc () jj==11[ ∞ 2 ==x,edx (Monotone Convergence Theorem) ∫ ∑ j c () [ j=1 = ∫ x 2 c ()dx . [

Hence we have not only shown that Cc ∈ S , but also proven that

trace ⎡⎤= x 2 dx ■ ⎣⎦Cc ∫ c () [

It’s possible for a probability measure to have a well defined correlation operator, while the mean value can not be defined. The next example illustrates the above,

∞ Example 4.11 : Let be a probability measure in such that tdt2 <∞ and c ∫ c x () −∞ ∞ tdt=∞. Then exists, but doesn’t exist. ■ ∫ cx () Cc mc −∞

72 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

3.5. Moments from the Characteristic Functional

In the theory of random variables we have seen that the characteristic function includes all the information of the stochastic process. As a result we can have all the moments of the random variable from the characteristic function. In this section we will see how we can take analogous results for the characteristic functional. Let Y be the characteristic functional of the probability measure c defined on [ . Let, also YV∈ 2 (i.e. the space of twice differentiable functionals). Calculating the first Frechet derivative we have

dxtz+ ⎛⎞ Y () d ⎜ ix+ tzy, ⎟ δYc()[]xz==⎜ ∫ e() dy⎟ = dt dt ⎝⎠⎜ ⎟ t=0 [ t=0 (1) =⋅izye∫ , ix, yc () dy [ Thus, we have,

δ 0,zi=⋅ zy dyimz =⋅ , ⇔ Yc()[] ∫ () c [

mzc ,0=− i ⋅δY[()[] z z ∈ (2)

Getting the second Frechet derivative we have,

dxtwzδ + ⎛⎞ 2 Y ()[]d ⎜ ix+ twy, ⎟ δ Yc()[xzw,, ]==⎜ i∫ zye() dy⎟ = dt dt ⎝⎠⎜ ⎟ t=0 [ t=0 =−∫ wy,, zyeix, yc () dy (3) [

Thus, we have,

δ 2 0,zw=− w ,y z ,y dy =− Czw , Yc()[ ] ∫ () c [ 2 Czwc ,0,,=−δ Y[()[ zw ] zw ∈ (4)

Similarly, we can get greater order quantities, depending on the smoothness of the functional.

Another very useful property of the characteristic functional is the ability to get moments, in the usual sense, for specific values of the time or spatial variable of the stochastic process. Let [ be the space of C1 (N , ) and ii, the corresponding inner product. Then, following the notation of Section 2.1.2/Chapter 2 we have ∞ EXtβ ⎡⎤;β ==⋅ mt xPdx ⎣⎦()X ()∫ Xt()() −∞ But from equation (2) we have

zy,,0 dy==−⋅∈ m z iδ z z ∫ c[() c Y ()[] [

Using Remark 2.5 we have

3.5. MOMENTS FROM THE CHARACTERISTIC FUNCTIONAL 73

∞ tdti=− ⋅δ 0 zz ∈ (5) ∫ cYz () ()[] [ −∞ where

cc[z ()ayzya=∈{};, ≤

Setting ztt=−δ( 0 ) we have

()ayytaPa=∈{}; ( ) ≤=() (6) cc[δ()tt− 0 0 Xt()

Thus,

δ β ⎡⎤βδδ=− ⋅ ⎡⎤ − =− ⋅ Y (7) EXt⎣⎦()00;0 iY ()() ⎣⎦ tt i () 0 δxt()0 δY where ()i denotes the Volterra derivative of the functional Y ()i at the point t . δxt() Remark 5.1 : Hence, we see that the mean value element is of the same kind as the Hilbert space. For a space of functions, the mean value element is also a function and the classical moments are integrals with respect to measures defined by the probability of the values at an instant time. ■

Example 5.2 : Consider, for example the space of continuous functions defined on the time interval []0,T , bounded above and below from the functions ctU ( ) and ctD ( ), i.e. the space of functions

h = {}uC∈≤≤∀∈()[]0, T : ctDU() utct () () , t [ 0, T ]

Then the mean value element mt( ) will be a function into the space h , as illustrated to the figure below. The mean value at time instants ti will be the value of the mean element, mt( i ) , or equivalently the mean value of the distribution connected with the values along the line εi .

ε2 ε1 x(t) ε3

h ctU ( )

mt( )

ctD ( )

0 t t1 t2 t3 T ■ For the general case we have

∞ nn n β ⎡⎤12 M nn12 nM EXt()12;ββ⋅⋅⋅=⋅⋅⋅ Xt () ; ... Xt (Mttt ; β ) x x… x PXt(),,… Xt ( )() dx ⎣⎦⎢⎥∫ 12 M 1 M −∞

74 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

where P is the joint probability distribution for the time instants tt,,… , t. Xt()1 ,,… Xt (M ) 12 M

Using Defintion 2.2 we will have ()ayytaytayta=∈{};, () ≤ ( ) ≤ ,… ( ) ≤ cc[δδ()(tt−−1 ,,… ttM ) 1122 MM = Pa() , a =∈()aa,,… a M Xt()1 ,,… Xt (M ) 12 M Using same arguments, as before we can show that

EXtβ ⎡⎤();ββnn12⋅⋅⋅= Xt () ; ... Xt ( ; β ) nM ⎣⎦⎢⎥12 M ⎡⎤

1 nn1++… M ⎢⎥ =−−−−=++… δδδδδY ()()()(0⎢⎥tt11 ,…… , tt ,..., ttMM ) , , ( tt ) (8) inn1 M ⎢⎥ ⎣⎦nn1 M 1 δ nn1++… M Y ()0 = inn1++… M nn12 nM δδxt()12⋅⋅⋅ xt ()… δ xt (M )

Again we have two consistent interpretations of the moments EXtβ ⎡ ();ββnn1 ⋅⋅ ... Xt ( ; )M ⎤ as ⎣⎢ 1 M ⎦⎥ in Example 5.2. More specifically, the above moment can be interpreted as the value of the corresponding operator at tt12,,… , tM or the moment of the joint distribution connected with the values along the lines εε12,,… εM .

Another important property of the characterization of a stochastic process through the characteristic functional or the probability measure is the possibility to get moments for the time derivatives of the stochastic process. From equation (5) we have ∞ tdti=− ⋅δ 0 zz ∈ ∫ cYz () ()[] [ −∞ where

cc[z ()ayzya=∈{};, ≤

Setting ztt=−δ′( 0 ) we have

()ayytaPa=∈{}; ′ ( ) ≤=() (9) cc[δ′′()tt− 0 0 Xt()

Thus Eβ ⎡⎤Xt′′;0βδδ=− i ⋅ ⎡⎤ t − t ⎣⎦()00Y ()() ⎣⎦ (10)

Similarly we can get general formulas

nn n ⎡⎤()mm12() () mM nn1++… M β ⎡⎤⎡⎤⎡⎤12 M i⋅⋅⋅⋅= E⎢⎥⎢⎥⎢⎥⎢⎥ Xt()11;ββ Xt() ; ... Xt() 1 ; β ⎣⎦⎢⎥⎣⎦⎣⎦⎣⎦ ⎡⎤ (11) ++… ⎢⎥()mmm() () () m =−−−−δδδδδnn1 M Y ()0⎢⎥11()tt ,…… ,() tt ,...,MM() tt , , () tt ⎢⎥11 MM ⎣⎦nn1 M

3.6. CHARACTERISTIC FUNCTIONS FROM THE CHARACTERISTIC FUNCTIONAL 75

3.6. Characteristic Functions from the Characteristic Functional

Another essential property of the characteristic functional is the ability to get characteristic functions, in the usual sense, for specific values of the time or spatial variable of the stochastic process. However more important is the possibility of calculating characteristic functions for the derivatives of the responces.

Using the definition of Section 2.1.3/Chapter 2. for the characteristic function, we have,

+∞ φυ≡=Eeβυix eP ix υ dx (1) Xt()() ()∫ Xt()() −∞ For the characteristic functional of the probability measure, we have, by definition,

Yc[()xe=∈∫ ix, y () dyx [

Setting x = υ z , where υ ∈ , and using Remark 2.5. we have

∞ υ ze==izyυ , dyedtzitυ ∈ (2) Ycc[()∫∫()z () [ −∞ where

cc[z ()ayzya=∈{};, ≤ , a ∈

Setting ztt=−δ( 0 ) we have ()ayytaPa=∈{}; ( ) ≤=() , a ∈ (3) cc[δ()tt− 0 0 Xt()0

Thus,

φυXt()()=−Y () υδ (tt0 ) , υ ∈ (4)

For higher dimensions we have analogous formulas. More specifically, by setting to the M characteristic functional x =++υυ11zz… M M , where υ =∈()υυ12,,… υM and M z =∈()zz12,,… zM [ we have

Y ()υυ11zz++… MM = izyzy{}υυ,,+++… υ zy , ==∫ edy11 2 2 MM c () (5) [ ∞ it{}υυ++ t… υ t = ed11 2 2 MM ()t ∫ c zz12,,,… zM −∞ where

()a =∈y ;,zy ≤ a ,, zy ≤ a ,,… z , y ≤ a (6) cc[zz12,,,… zM {} 1 1 2 2 MM

Setting zttii=−δ(), iM=1,2,… , and using Definition 2.2 we have

()a=∈{} y;, yt () ≤ a yt ( ) ≤ a ,… yt ( ) ≤= a cc[δδ()(tt−−1 ,,… ttM ) 1122 MM

76 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

= Pa() , a =∈()aa,,… a M Xt()1 ,,… Xt (M ) 12 M Hence, φυδυδ()υ =−++−() (tt )… ( tt ) , υ =∈()υυ,,… υ M (7) Xt()1 ,,… Xt (M ) Y 11 MM 12 M

Additionally we can calculate characteristic functions for the time derivatives of the stochastic process. Setting to equations (5), (6) zttii=−δ′(), iM=1, 2,… , and using Definition 2.2 we have

()a=∈{} y;, yt′′ () ≤ ayt ( ) ≤ a ,… yt ′ ( ) ≤= a cc[δδ′′()tt−−1 ,,… ( ttM ) 112 2 MM

= Pa() , a =∈()aa,,… a M Xt′′()1 ,,… Xt (M ) 12 M Hence,

φυδυδ()υ =−++−()′′ (tt )… ( tt ), υ =∈()υυ,,… υ M (8) Xt′′()1 ,,… Xt (M ) Y 11 MM 12 M

Finally, we can prove the general formulae, in the same way as before. We have from

(ni ) Definition 2.2 and equation (5) for zttii=−δ (),

(nn) ( ) ( n) ayytaytayta=∈;,12 ≤ ≤ ,… M ≤= cc[()nn1 ()M () {}()11() 2 2 ()MM δδ(tt−−1),,… ( ttM )

= Pa() , a =∈()aa,,… a M Xt′′()1 ,,… Xt (M ) 12 M

Hence we get the equality,

(nn) ( ) φυδυδυ =−++−1 tt… M tt , υ ∈ M (9) ()nn1 ()M () Y ()11() MM() Xt( 1),,… X( tM )

Thus we have general expressions for the calculation of characteristic functions from the characteristic functionals for the values of the stochastic process as well as for the derivatives.

Remark 6.1 : It is very useful for applications to work in the level of characteristic functional. Then all the above results gives us a method to interpret or visualize the resulting characteristic functional by taking the characteristic functions for finite dimensional subspaces and then calculate the Fourier transform to get the corresponding probability density functions. ■

3.7. GAUSSIAN MEASURE 77

3.7. Gaussian Measure

Gaussian measures plays a very important role in the theory of probability measures in infinite dimensional spaces. Additionally, as we will see at Chapter 5, it is an essential tool, for the representation of the responses of stochastic dynamical systems. Their practical importance basically relies to the fact that we have analytical expressions for the infinite dimensional integrals of cylinder functions with respect to Gaussian measures. Another important feature is the localization of the probability mass into the infinite dimensional space of functions. In other words Gaussian measures have the property of kernel characteristic functionals, that it wll be discussed extensively at Chapter 5/Section 5.4. The last property in combination with the property of analytical computation of integrals gives us a very promising tool for the numerical solution of functional differential equations, discussed in Chapter 5. Hence its very essential to study extensively Gaussian measures in [ . We start with the following

Definition 7.1 : A Gaussian measure cG in [ is a probability measure in [ such that for each x ∈[ , the measurable function x,i is normally distributed, i.e. there exists real numbers mx and σx such that

2 a ()tm− 1 − x ayxyaedt=∈;, ≤= 2σx ■ (1) cc[x () G {}∫ −∞ 2πσx

We will now compute the mean and covariant operator for the Gaussian measure. We have,

mx,,=∈ zx dzx c ∫ c[G () [ Using Remark 2.5 the above integral turns out, to be

2 ∞∞()tm− − x 1 σ mx, ==⋅= t dt t e2 x dtm (2) c ∫∫c x () x −∞ −∞ 2πσx

For the covariance operator we will have

Cxx,,=∈ xz2 dz x c ∫ c[G () [

Using Remark 2.5, we will have

2 ∞∞()tm− − x 1 σ C x, x==⋅= t22 dt t e2 x dt σ (3) c ∫∫cx () x −∞ −∞ 2πσx

Concerning the characteristic functional of a Gaussian measure cG we have the following

Lemma 7.2 : Let cG be a Gaussian measure in [ . Then its characteristic functional is given by Cxx, im, x − c c 2 Y G ()xe= (4)

78 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

where mc is the mean of cG and Cc the covariance operator of cG . Proof : Using Remark 2.5 we have

x ==edyedtix, y it YccGGx() ∫∫() () [[

where cc[xG()ayxya=∈{};, ≤ is the distribution of x,i . Hence,

()tm− 2 − x it 1 2σx Y G ()x = eedt ∫ πσ [ 2 x Make a change of variables first and use contour integration to conclude

σ im − x x 2 Y G ()xe= .

But from (2) - (3)we have

mmxx = c , and σx = Cxxc ,

Hence

Cxx, im, x − c c 2 Y G ()xe= ■

Remark 7.3 : Since a probability measure is uniquely defined by its characteristic functional, then, in view of the above lemma, a Gaussian measure is uniquely defined by its mean and covariance operator. ■

For the characterization of a Gaussian measure we have the following Cxx, ix, x− c 0 2 Theorem 7.4 [Prohorov] : Let x0 ∈[ and Cc ∈ S . Then Y ()xe= is the characteristic functional of a Gaussian measure in [ .

Proof : Clearly Y ()01= and Y is positive definite. Consider first the case x0 = 0 . Then

Cxx, − c 1 1Re−≤−≤Y ()x 1eCxx2 , 2 c because 1−≤ey−y for all y ≥ 0 . 1 Since C ∈ S part (b) of Minlos-Sazonov Theorem. (Theorem 3.7) is trivially satisfied. 2 c Therefore by the conclusion of the Theorem, there exists a Borel measure c such that

Yc[()xe=∈∫ ix, y () dyx [

That is,

Cxx, − c ∫ edyexix, yc[()=∈2 [

3.7. GAUSSIAN MEASURE 79

Let cc[x ()ayxya=∈{};, ≤ be the distribution of x,i . Then using Remark 2.4 we have

Cxx, − c edteit = 2 ∫ c x () [

Cxx, Hence c x is a normal distribution with mean 0 and variance c . That is c is a Gaussian measure in [ . Consider now the case x0 ≠ 0 . Let

Cxx, − c Y()xe= 2

Then

YY()x = exix0 , x ()

By the case already proved, there exists a Gaussian measure c in [ such that

Yc[()xe=∈∫ ix, y () dyx [ Define a Borel measure c in [ as follows

cc()EExE=− (0 ), ∈ U[()

It is easy to see that c is a Gaussian measure and

∫ edyexxix, ycYY()==ix0 , x () () ■ [

Finally, as we mention at the beginning, analytical integration of cylinder functions over Gaussian measures can be carried out. Hence for Gaussians measures we have

g x,,, y……… x , y dy== g t ,, t dt ,, dt ∫∫()111nnncc[ () ( )x ( ) [ n −n /2 −1/2 ⎪⎪⎧⎫1 T =−−−()2detπ ()Cg()uumumu exp⎨⎬⎪⎪[][] C−1 d. ∫ ⎪⎪2 () n ⎩⎭⎪⎪

Where, C−1 is the inverse for matrix ⎡CCxx ⎤ = , , ij,= 1,… , n, m= mx,,,,… mx . ⎣⎢ ij⎦⎥ i j ()1 n m,i is the mean value of the measure and Cii, is the correlation functional of the measure. In Chapter 5/Section 5.1 we will give the precise formulation as well as the proof of the above result.

80 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

3.8. Tatarskii Characteristic Functional

As we saw at the previous section the characteristic functional for the Gaussian measure can be represented by explicit formulae. There exist few other cases for which it is possible to present such formulas. But the problem of extending the class of random functions for which it is possible to write down the characteristic functional explicitly is important for many applications. If we posses an explicit formula for the characteristic functional that contains several arbitrary functions, it is possible to use this model for an approximate solution of different statistical problems. In this section we will present a characteristic functional representation, first introduced by V.I. Tatarskii (1995). This general expression creates a wide class of characteristic functionals. The above representation is based on the construction of a random function, which we will describe in the sequel. Let the random function depending on vector argument r ∈ n of the form

⎧0 with probability p ⎪ 0 ⎪ ⎪ξ11gp()rr, with probability 1 ⎪ ⎪ ⎪ X ()r ;β = ⎨⎪ (1) ⎪ ⎪ N ⎪ ξ gprr, with probability ⎪∑ kk() N ⎪ k=0 ⎪ ⎩⎪

where the numbers pN obey the conditions

pN ≥ 0 ∑ pN =1 (2) N=0

We assume the characteristic function

∞ χυ()= ∑ pN exp (iN υ ) (3) N=0 is known. The same information is available from the function

∞ NN Ψ≡()AA =∑ pAiAN =−χ()log (4) N=0

The function Ψ()A , the generating function of factorial moments, is related to the moments of random N by the formula

dAn Ψ() =−−−+NN()()(12 N… N n 1 ) ANn− (5) dAn

Thus, dAn Ψ() =−−−+≡NN()()(12 N… N n 1 ) F (6) dAn n A=1 3.8. THE TATARSKII REPRESENTATION OF A CHARACTERISTIC FUNCTIONAL 81

The numbers Fn are called the factorial moments of the random variable N . The amplitudes

ξk are assumed to be statistically independent and have the same probability distribution and characteristic function independent of k :

ϕυ()= exp (i υξk ) (7)

We assume that all rk are statistically independent and have the same probability density function W ()rk .Let us consider the characteristic functional for X ()r ;β given by (2). Substituting (1) into (2) we obtain

⎡ N ⎤ ()x = exp ⎢ixdξ g()rr, () r r⎥ (8) Y ⎢ ∑ k ∫ k ⎥ ⎣ k=0 ⎦

We are able to average over ξ , and using (7) we obtain

N = ϕ g rr, r r Y ()x ∏ (∫ ()k xd() ) (9) k=0

Here, N and rk remain random. Because all rk are statistically independent and have the same probability density function W ()rk , averaging over rk results in

N = rrϕ g rrrr, Y ()x ∏∫∫dWkk ( ) ( () k x() d) (10) k=0

All factors in this product differ only by the notation for the variable of integration and are equal to each other. Thus,

⎡⎤N Y ()xdW= ⎢⎥rr′′()ϕ g() rrrr, ′ xd() (11) ⎣⎦∫∫( )

To fulfil the averaging with respect to N , we can use (4) with the value A determined by the expression

AdW= ∫∫rr′′()ϕ( g() rrrr, ′ xd() ) (12)

Substituting (12) into (11) we obtain

Y ()x =Ψ(∫∫dWrr′′()ϕ( g() rrrr, ′ x() d)) (13)

This lead us to the following

Theorem 8.1 : Let n a) W : → + , be an arbitrary non-negative function, that is normalized to 1 b) ϕυ() be an arbitrary characteristic function c) g()ii, : nn×→ a completely arbitrary function and d) Ψ ()A be an arbitrary generating function of factorial moments, i.e. be of the form Ψχ()AiA=− (log ), where χ is an arbitrary characteristic function. Then the functional

82 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

Y ()x = Ψϕ(∫∫dWrr′′()( g() rrrr, ′ x() d)) is a characteristic functional. ■

Remark 8.2 : Although the proof of the above theorem was carried out for the case of r ∈ n being a spatial variable the same results hold for the case when we substitute r with t being time variable. ■

For the calculations that follow we will assume that the functions are defined onto the time interval I =[]0,T .

3.8.1. Mean and Correlation Operators

As we compute at Section 4 of the present chapter the mean value operator of a probability measure is

mzc ,0=− i ⋅δY[()[] z z ∈

By direct calculation we can find that the mean operator will have the form

mz,10=− i ⋅Ψ′′ ⋅ϕ ⋅ Wt ′gtt, ′ ztdtdtz ′ ∈ (14) c () ( )(∫ ∫ () ()() ) [

Using the above result and equation (7) we will have that

EXtβ ⎡⎤;10βϕ=− i ⋅Ψ′′ ⋅ ⋅ Wtgt, t dt (15) ⎣⎦(0 ) () ( )(∫ () ()0 )

Concerning the covariance operator, we have proved from Section 4 that

2 Czwc ,0,,=−δ Y[()[ zw] zw ∈

So, by calculating the second Frechet derivative, we will wave

2 2 δϕ2 ()[0,z z ]=Ψ′′ () 1 ⋅⎡⎤ ′ () 0 ⋅⎡ W() t ′gtt(), ′ z() t dt dt ′⎤ + Y 12 ⎣⎦∏⎣⎢∫∫ k ⎦⎥ k=1 (16) ⎛⎞2 +Ψ′′′10 ⋅ϕ ⋅⎜ W t ′ ⋅ gtt, ′ z t dt dt ′⎟ () ( )⎜∫∫()∏()()k () ⎟ ⎝⎠k=1

3.8.2. Characteristic functions

As we compute at Section 6 the joint characteristic function at time instants tt12,,… , tM of a probability measure is

φυδυδ()υ =−++−() (tt )… ( tt ) , υ =∈()υυ,,… υ M Xt()1 ,,… Xt (M ) Y 11 MM 12 M

By direct calculation we can find that the joint characteristic function will have the form

φυδυδ()υ =−++−=() (tt )… ( tt ) Xt()1 ,,… Xt (M ) Y 11 MM , υ ∈ M (17) =ΨWtϕυgt,, t +… + υ gt t dt (∫ () ()1 ()1 M (M )) 3.8. THE TATARSKII REPRESENTATION OF A CHARACTERISTIC FUNCTIONAL 83

For characteristic functions concerning time derivatives a of the stochastic process we have using equation (9)

(nn1) ( M ) φυδυδ()nn() ()υ =−++−=()tt… () tt Xt1 (),,… XM () t Y ()11 MM 1 M , υ ∈ M (18) nn1 M =ΨWt()ϕυ ∂gt()1 ,, t +… + υ ∂ gt (M t ) dt (∫ ()1 tMt1 M )

3.8.3. The 2D Tatarskii Representation of a Characteristic Functional

For applications of characteristic functionals connected with stochastic differential equations, it is usually needed to have representations of vector valued stochastic processes. In this section we generalize the Tatarskii representation for the two dimensional case. Based on the same arguments we can easily extend the generalization presented above to the n− dimensional case. We consider the random function depending on the n-dimensional vector argument r of the form ⎪⎧ with probability ⎪()0,0 p0 ⎪ ⎪⎛⎞′′′⎟ ⎪⎜ξ11g rr, ,ζ 11hprr, ⎟ with probability 1 ⎪⎜⎝⎠()()⎟ ⎪ ⎪ X ()r ;β = ⎨⎪ (19) ⎪ ⎪ ⎪⎛⎞NN ⎪⎜ ξ gh()rr,′′′, ζ () rr ,⎟ with probability p ⎪⎜∑∑kk kk⎟ N ⎪⎝⎠kk==00 ⎪ ⎩⎪

(there is a chance to generalize further by assuming different N and M for every component, but the things gets very complicated. For our purpose the above generalization is the most suitable.)

The numbers pN obey the conditions

∞ pN ≥ 0 ∑ pN =1 (20) NM,0=

We assume the characteristic function

∞ χυ()= ∑ pN exp (iN υ ) (21) NM,0= is known. The same information is available from the function

∞ NN Ψ≡()AA =∑ pAN (22) NM,0=

The r.v.’s ()ξζkk, are assumed to be statistically independent and have probability distribution and characteristic function independent of k :

ϕυν(),exp=+()i [ υξkk νζ ] (23)

84 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

We assume that all rr′′′, are statistically independent and have the same probability density ()kk function W rr′′′, . Let us consider the characteristic functional for X r ;β given by (2). ()kk () Substituting (19) into (2) we obtain

⎡⎤NN ()x,expyi=+⎢⎥ξζg rr,,′ xdhyd() r r rr′′ () r r (24) Y ⎢⎥∑∑kk∫∫()kk() ⎣⎦kk==00

We are able to average over ξζ , , and using (23) we obtain

N = ϕ g rr,,′′′ rr rr rr Y ()x,,yxdhyd∏ (∫∫()kk() ()() ) (25) k=0

′′′ ′′′ Here, NM, and rrkk , remain random. Because all rrkk , are statistically independent and have the same probability density functionW rr′′′, , averaging over r ′ and r ′′ results in ()kk k k

N xy,,= Wrr′′′′ϕ g rr,, x r d r , h rr ′′ y r d r d r ′ d r ′′ (26) Y ()∏∫∫()kkk ( ()() ∫() kkk() ) k=0 All factors in this product differ only by the notation for the variable of integration and are equal to each other. Thus,

⎡⎤N Y ()xy,,= ⎢⎥ Wrr′′′′ϕ g() rr,, x() rr d , h() rr ′′ y() rrrr d d ′ d ′′ (27) ⎣⎦⎢⎥∫∫()kkk ( ∫ kkk)

To fulfil the averaging with respect to N , we can use (22) with the value A determined by the expression ⎛⎞ ⎜ ⎟ AW= r′′,, r′′ϕ⎜ g() rr,, xdhyddd() r r() rr ′′() r r⎟ r ′ r ′′ ∫∫()⎜ ∫ ∫ ⎟ rr′′ ′ ⎝⎠ r r Substituting we obtain

⎛⎞⎛⎞ ⎜ ⎜ ⎟ ⎟ Y ()x,,yW=Ψ⎜ rr′′′′ϕ⎜ g() rr,, xdhyddd() rr ,() rr ′′() rrrr⎟ ′ ′′ ⎟ (28) ⎜∫∫()⎜ ∫ ∫ ⎟ ⎟ ⎝⎠⎜ rr′′ ′ ⎝⎠ r r ⎟

This formula represents the characteristic functional for a random function of the type (19) Analogous generalizations for the n−dimensional case can be carried out leading to results of the form

⎡⎤ ⎛⎞⎟ x =Ψ⎢⎥ ⋅⋅ Wxdgxdddr,,……… rϕ⎜ g rr,, rr ,, rr rrr⎟ r Y ()⎢⎥∫∫ (111NNNN )⎜ ∫()1 () ∫ ()N () ⎟ 1 ⎢⎥⎝⎠⎜ ⎟ ⎣⎦rr1 N r r with x ∈××[[… . N

3.9. REDUCTION OF THE TATARSKII CHARACTERISTIC FUNCTIONAL 85

3.9. Characteristic Functionals derived from Tatarskii Functional

As we saw in the previous section the Tatarskii characteristic functional represents a wide class of stochastic processes. In what follows we will study how the general representation can be reduced to characteristic functions corresponding to useful measures for applications. These measures will be used at Chapter 5/Section 4 for the representation of the responses of the stochastic dynamical systems. These specific representations are very useful for this aim since Gaussian measures are very restricted, while, Tatarskii characteristic functional is very impractical for calculations, since it involves many unknown functions. The analysis presented below will follow EGOROV, A.D. & SOBOLEVSKY, P.I. & YANOVICH, L.A. (Functional Integrals: Approximate Evaluation and Applications). Consider the special form of Tatarksii characteristic functional with

Ψ ()A = exp {}A (1) ∞ inσ ϕ()x = ∑ n xn (2) n=1 n! and the function g()rr, ′ depends only from the first argument r , i.e.

g()rr, ′ = l() r (3)

Based on the above assumptions the characteristic functional takes the form

⎧⎫∞ n ⎪⎪i σn Y ()x = exp⎨⎬Klx() , ,… , lx , (4) ⎪⎪∑ n ⎩⎭⎪⎪n=1 n!

where Klxnn( 1,,,,… lx) is a real symmetric continuous n− linear form on [[××… n times and σn ()n =1, 2,… are real and complex parameters. Some σn may be equal to zero here, and the numbers of summands may be finite. Let us now consider some more restricted classes of functionals of form (4). Let h be a set with a finite measure d defined on it, a ∈[ , ρ(u) is a mapping from h into [ and parameters σn are real. Then we may define

∞ n ⎪⎪⎧⎫n ⎪⎪i σn Yd()x =+exp⎨⎬iax ,ρ() u , x() du (5) ⎪⎪∑ n! ∫ ⎩⎭⎪⎪n=1 h

A possible example of such functional is

⎧⎫ ⎪⎪ Yd()x =+exp⎨⎬iax , gρ() u , x() du (6) ⎪⎪∫ () ⎩⎭⎪⎪h where gz() is a function which is analytic in the vicinity of zero ((6) is reduced to (5), if we d n set σn = nngz() , σσ01==0 ). The following two parameter family of functionals idz z=0

86 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

⎪⎪⎧⎫ ⎪⎪11iuxαρ(),, 1 iux βρ() Yd()x =+exp⎨⎬iax , e −−− 1 e 1 () du (7) ⎪⎪αβ− ∫ α() β() ⎩⎭⎪⎪h where αβ, are real parameters, αβ≥ , is the special case of (5). If β → 0 , we obtain the functional

⎪⎪⎧⎫ ⎪⎪1 ⎡⎤iuxαρ(), Yd()x =+exp⎨⎬iax , e −− 1 iαρ() u , x() du (8) ⎪⎪α2 ∫ ⎣⎦⎢⎥ ⎩⎭⎪⎪h which corresponds to Poisson measure. If α → 0 and β → 0 , then we obtain the functional

⎧⎫ ⎪⎪1 2 Yd()x =−exp⎨⎬iax ,ρ() u , x() du (9) ⎪⎪2 ∫ ⎩⎭⎪⎪h which corresponds to Gaussian measure. As we see at previous sections, the characteristic functional of a Gaussian measure has the form

⎪⎪⎧⎫1 Y ()x =−exp⎨⎬imx , Kxx() , (10) ⎩⎭⎪⎪2 where Kxx(), is an positive definite bilinear one. One more example of functionals of type (6)

⎧⎫ ⎪⎪ Yd()x =−−+exp⎨⎬iax ,⎡⎤ ln 1 iσρ() u , x i σρ() u , x() du (11) ⎪⎪∫ ⎣⎦⎢⎥() ⎩⎭⎪⎪h

We would like to mention separately the class when [ is some set of functions x(t) on segment []0,T ⊂ . Then

T T ax, = at xtdt , au,ρρ= uatdt (12) ∫ () () () ∫ t ()() 0 0

where the functions at( ), ρt (u) may also be the generalized ones. In this case a more special form of the Poisson measure is for

α =1, ⎪⎧1, iftuT∈[] , ρ ()u = ⎨⎪ , t ⎪ ⎩⎪0 , otherwise and

h =[]0,T with d ()du=⋅λ du , λ ∈

Then

TT ∫∫∫ρ()u,, xd () du== x () t dtdu a∗ x h 0 u

Setting aa= ∗ we will get

⎪⎪⎧⎫ ⎪⎪⎡⎤iuxαρt (), Y ()x =−exp⎨⎬λ edu 1 (13) ⎪⎪∫ ⎣⎦⎢⎥ ⎩⎭⎪⎪T 3.9. REDUCTION OF THE TATARSKII CHARACTERISTIC FUNCTIONAL 87

which is a simpler version of the general Poisson characteristic functional of equation (8). Another simplification, for the case when [ is some set of functions x(t) on segment []0,T ⊂ can be made to equation (11) by choosing again as ⎪⎧1, iftuT∈[] , ρ ()u = ⎨⎪ , t ⎪ ⎩⎪0 , otherwise and

h =[]0,T with d ()du= du , Choosing a in a suitable way we can get the Gamma characteristic functional for the infinite dimensional Gamma distribution ⎧⎫ ⎪⎪ Y ()x =−exp⎨⎬⎡⎤ ln 1iuxduσρ () , (14) ⎪⎪∫ ⎣⎦⎢⎥()t ⎩⎭⎪⎪T Finally, for the case of [ being set of functions x(t) on segment []0,T ⊂ functional (6) may be considered as the characteristic functional of a random process X ()t which may represented as follows,

X tuduattT=+∈ρ ,0, (15) ()∫ t ( )j ( ) () [ ] h where j is a stochastic orthogonal measure,

β ⎡⎤⎡⎤ EiA()exp⎣⎦⎣⎦λλjd()= exp gA() () , A ⊂h 2 β = σ ⊂ E ()jd()AA2 () , A h

Another interesting class of functionals is of form (6) with

biz2 ⎛⎞izλ λ gz()=− iaz z +⎜ e −−1 ⎟πλ() d (16) ∫ ⎝⎠⎜ +λ2 ⎟ 21 where a ∈ , b ≥ 0 , πλ()ddM= () λ for λ < 0 and πλ(ddN)= ( λ) for λ > 0 ; M (λ), N ()λ satisfy the following conditions 1. M (λ) and N ()λ are nondecreasing functions on ()−∞,0 and ()0,∞ , respectively, 2. MN()−∞ = () ∞ = 0 , 0 ε 3. ∫ λλ2dM ()<∞ , ∫ λλ2dN ()<∞ for any ε > 0. −ε 0

If h =[]0,T , d (du)= du , then bearing in mind (12), we obtain

88 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

⎧ TT TT ⎪ b Y ()x=−+exp⎨ iaρ ()() u x s dsdu B()()() t , s x t x s dsdt ⎪ ∫∫s 2 ∫∫ ⎩⎪ 00 00 (17) TT⎡⎤⎧⎫ T ⎪⎫ ⎪⎪iλ ⎪ +−−⎢⎥exp⎨⎬iuxsdsλρ()() 1 ρ()() uxsdsdud πλ ( ) ⎬⎪ ∫∫⎢⎥⎪⎪ ∫ss1+λ2 ∫ ⎪ 00⎣⎢⎥⎩⎭⎪⎪ 0 ⎦ ⎭⎪

T where B ts, = σρ u ρ udu, σλπλ=+bd2 . This functional is the characteristic ( )2 ∫ st () () 2 ∫ () 0 one of the random process

T X tsd= ρξ , (18) ()∫ ts () 0

where ξs is a homogeneous process with independent increments which satisfies the condition

ξ0 = 0 . Note that (17) for

MN()λλ== () 0 gives rise to functionals of type (9); it gives rise to functionals of type (8) for

⎪⎧−≤a ifλ 1 b = 0 , M ()λ = 0 , N ()λ = ⎨⎪ ⎩⎪ 0otherwise and to functionals of type (11) for

∞ 1 ⎛⎞υ b = 0 , M ()λ = 0 , Nd()λυα=−exp⎜ −⎟ , > 0 . ∫ υα⎝⎠⎜ ⎟ −λ

We shall mention still another characteristic functional which corresponds to the measure which we shall call the Abel measure,

⎧⎫−1 ⎪⎪⎛⎞1 Y ()xKxx=+exp⎨⎬⎜ 1() , ⎟ (19) ⎪⎪⎝⎠⎜ ⎟ ⎩⎭⎪⎪2

The characteristic functional of the Cauchy measure,

⎧⎫∞ ⎪⎪ Y ()x =−exp⎨⎬axe , , (20) ⎪⎪∑ jj ⎩⎭⎪⎪j=1 ∞ where a j are positive numbers, ∑ a j <∞, {}e j , j =1,2,…, is a basis in [ , is an example j=1 of the characteristic functional which cannot be represented in form (4). Characteristic functional of form (4) give examples of measures, when the number of summands in the exponent is finite, the last summand has the index np= 2 , p >1 and p+1 2 p σσ2 p =−()1 . In particular let us mention the case, when

3.9. REDUCTION OF THE TATARSKII CHARACTERISTIC FUNCTIONAL 89

⎧⎫2 p ⎪⎪σ 2 p Yd()x =−exp⎨⎬ρ()ux , () du , (21) ⎪⎪()2!p ∫ ⎩⎭⎪⎪h where the notation is the same as for (5). For the space of functions on []0,T , (21) will assume the form,

2 p ⎪⎪⎧⎫2 p ⎡⎤T ⎪⎪σ Y ()x =−exp⎨⎬⎢⎥ρ ()()uxtdt du , (22) ⎪⎪()2!p ∫∫⎢⎥t ⎩⎭⎪⎪h ⎣⎦⎢⎥0

Finally, when there is a finite series in the exponent of (4) and the last summand enters with the imaginary term then we obtain pseudomeasures (see EGOROV, A.D. & SOBOLEVSKY, P.I. & YANOVICH, L.A. , Functional Integrals: Approximate Evaluation and Applications [page.5] for details), namely, the Feynman measures. For example,

⎧⎫ ⎪⎪i 2 Yd()x =−exp⎨⎬ρ()ux , () du , (23) ⎪⎪2 ∫ ⎩⎭⎪⎪h in particular,

⎪⎪⎧⎫⎡⎤T 2 ⎪⎪i Yd()x =−exp⎨⎬⎢⎥ρ ()()uxtdt ( du ) . (24) ⎪⎪2 ∫∫⎢⎥t ⎩⎭⎪⎪h ⎣⎦⎢⎥0

90 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

3.10. References

ASH, R.B., 1972, Measure, Integration and Functional Analysis. Academic Press. ASH, R.B., 2000, Probability and Measure Theory. Academic Press. BERAN M.J., 1968, Statistical Continuum . Interscience Publishers. DORDNOV, A.A., 1977, Approximate Computation of Characteristic Functionals of Some Classes of Probability Measures. Issledovaniya po Prikladnoi Matematike 4, 83-91. EGOROV, A.D. & SOBOLEVSKY, P.I. & YANOVICH, L.A., 1993, Functional Integrals: Approximate Evaluation and Applications. Kluwer Academic Publishers. FOMIN, S.V., 1970, Some New Problems and Results in Nonlinear Functional Analysis. Mathematika 25, 57-65. FRIEDRICHS, K.O. & SHAPIRO. H.N. et al., 1976, Integration of Functionals. Courant Institute of Mathematical Sciences. FRISTEDT, B. & GRAY. L., 1997, A Modern Approach to Probability Theory. Birkhäuser. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1996, Theory of Random Processes. Dover Publications. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1974, The Theory of Stochastic Processes I. Spinger. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1975, The Theory of Stochastic Processes II. Spinger. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1975, The Theory of Stochastic Processes III. Spinger. HOPF, E., 1952, Statistical Hydromechanics and Functional Calculus. Journal of Rational Mechanics and Analysis 1, 87-123. KANDELAKI, N.P. & SAZONOV, V.V., 1964, On Central Limit Theorem for Random Elements with Values in Hilbert Spaces. Theory of Probability and Its Applications, IX, 38-46. KLIMOV, G., 1986, Probability Theory and Mathematical Statistics, MIR Publishers. KOLMOGOROV, A.N., 1956, Foundation of the Theory of Probability. KOTULSKI, Z., 1989, Equations for the Characteristic Functional and Moments of the Stochastic Evolutions with an Application. SIAM J. APPL. MATH., 49, 296-313. KOVALENKO, I.N. & KUZNETSOV, N.YU. & SHURENKOV, V.M., 1996, Models of Random Processes. CRC Press. KUO, H.H., 1975, Gaussian Measures in Banach Spaces, Springer. LANGOUCHE, F. & ROEKAERTS, D. & TIRAPEGUI, E., 1979, Functional Integral Methods for Stochastic Fields. Physica, 95A, 252-274. MONIN. A.S. & YAGLOM, A.M., 1965, Statistical Fluid Mechanics: Mechanics of Turbulence. MIT Press. PARTHASARATHY, K.R., 1967, Probability Measures in Metric Spaces. Academic Press. PFEFFER, W.F., 1977, Integrals and Measures. Marcel Dekker, INC. PROHOROV, YU.V., 1956, Convergence of Random Processes and Limit Theorems in Probability Theory. Theory of Probability and Its Applications, I, 157-214. PROHOROV, YU.V. & ROZANOV, YU.A., 1969, Probability Theory. Springer. PROHOROV, YU.V. & SAZONOV, V.V., 1961, Some Results Associated with Bochner’s Thoerem. Theory of Probability and Its Applications, VI, 82-87. PUGACHEV, V.S. & SINITSYN , 1987, Stochastic Differential Systems. John Wiley &Sons. 3.10. REFERENCES 91

ROSEN, G., 1967, Functional Integration Theory for Incompressible Fluid Turbulence. The of Fluids, 10, 2614-2619. ROSEN, G. & OKOLOWSKI, J.A. & ECKSTUT, G., 1969, Functional Integration Theory for Incompressible Fluid Turbulence. II. Journal of , 10, 415-421. ROSEN, G., 1960, Turbulence Theory and Functional Integration I. The Physics of Fluids, 3, 519-524. ROSEN, G., 1960, Turbulence Theory and Functional Integration II. The Physics of Fluids, 3, 525-528. SAZONOV, V.V., 1958, A Remark on The Characteristic Functional. Theory of Probability and Its Applications, III, 188-192. SIMON, B., 1979, Functional Integration and Quantum Physics. Academic Press. SKOROKHOD, A.V., 1974, Integration in Hilbert Spaces. Spinger. SKOROKHOD, A.V., 1984, Random Linear Operators. D. Reidel Publishing Company. SOBCZYK, K., 1991, Stochastic Differential Equations. Kluwer Academic Publishers. SOIZE, C., 1994, The Fokker-Planck Equation and its Explicit Steady State Solutions. World Scientific. TATARSKII, V.I., 1995, Characteristic Functionals for one Class of Non-Gaussian random Functions. Waves in Random Media, 5, 243-252. TESTA, F.J. & ROSEN, G., 1974, Theory for Incompressible Fluid Flow. Journal of The Franklin Insitute, 297, 127-133. VISHIK, M.J. AND FURSIKOV, A.V., 1980, Mathematical Problems of Statistical Hydromechanics. Kluwer Academic Publishers. VOLTERRA, V., 1959, Theory of Functionals and of Integral and Integro-Differential Equations. Dover Publications. VULIKH, B.Z., 1976, A Brief Course in the Theory of Functions of a Real Variable. MIR Publishers. WÓDKIEWICZ, K., 1981, Functional Representation of Non-Markovian Probability Distribution in Statistical Mechanics. Physics Letters 84A, 56-58.

92 CHAPTER 3 PROBABILITY MEASURES AND THEIR CHARACTERISTIC FUNCTIONALS IN HILBERT SPACES

Chapter 4

Stochastic Calculus – Principles and Results

In various problems of mathematical physics, many physical quantities are represented using stochastic processes. In the context of mathematical modelling we come naturally to concepts such us, continuity, derivatives and integrals of the functions that represent these physical quantities. It is therefore essential to define the above concepts for stochastic processes. In this direction three kinds of convergence have been developed, namely

• Mean Square Convergence • Convergence in Probability • Almost Sure Convergence

The definitions of the above are immediate generalizations of the convergence of sequences of random variables, given at Chapter 1/Section 1.3.5. In this chapter we will present the basic principles and results of mean square calculus. Using this calculus we reduce the problem of differentiation or integration of a stochastic process to the differentiation or integration of the covariance function. In Section 4.1.1-4.1.4 we present some basic definitions and results concerning the classical notions of m.s.-convergence, continuity and differentiation of second-order stochastic processes. Sections 4.1.5 deals with the issue of m.s.-integration of a stochastic process with respect to another (independent) process and Section 4.1.6 for the case of integration over martingales with non-anticipative dependence. Section 4.2 includes some fundamental results associated with analytical properties of sample functions, such as continuity, differentiation and integration. In what follows we assume the existence of a probability space()Ω, Uc(Ω), . The stochastic process X :Ω→CI() is defined on it, and the induced probability space is ()CI(),,Uc() CI () CI() .

93 94 CHAPTER 4 STOCHASTIC CALCULUS – PRINCIPLES AND RESULTS

4.1. Process of 2nd order – Mean-Square Calculus

4.1.1. Preliminaries

Definition 1.1 : A stochastic process X ()t ;β , tI∈ is called a second-order or Hilbert stochastic process, iff

EXtβ ()(); β 2 < ∞ We shall denote this space of stochastic processes as m (I). ■ (1)

The above definition can easily be extended for stochastic processes taking complex or vector values.

Remark 1.2 : The Schwarz inequality (Chapter 1/Section 1.3.2) implies that a second order stochastic process always has a covariance function. ■

Theorem 1.3 [Completeness property] : For a sequence {}X n of stochastic processes in m (I) to converge in m (I) to some limit, it is necessary and sufficient that the sequence

{}X n is fundamental. In other words the space m (I) is complete. Proof: The necessity of the above theorem is general for metric spaces as a consequence of the triangle inequality. Let us prove the sufficiency. Since EXββ22−≤− EX E β X X2 ()nn( ′′) ()[] nn β 2 if a sequence is a fundamental sequence, the sequence {}X n is bounded; thus EX()n < c. Furthermore the fact that the sequence is fundamental in m (I) implies that it is fundamental in measure. Consequently (See GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1996, Theory of Random Processes) the sequence X contains a subsequence X that converges almost uniformly {}n {}nk in . Using Fatou’s lemma, we obtain

cEXEXEX22≥≥=limββ lim 22 β ()nnkk()() from which it follows that X ∈m ()I . Let ε denote an arbitrary positive number. Since the sequence {}X n is fundamental, it follows that for all nn≥ 0 ,

22⎛⎞ 2 ε ≥−≥−=−limEXββ⎡⎤ X E⎜ lim ⎡⎤ X X⎟ EXX β[] ()⎢⎥nnkk⎜ ⎢⎥ nn⎟ () n kk→∞⎣⎦⎝⎠ →∞ ⎣⎦

which proves the convergence of the sequence {}X n to X ∈m ()I . ■

4.1.2 Mean-Square Convergence

Definition 1.4 : Let X ()t ;β , tI∈ be a second-order stochastic process. A process X ()t ;β converges in the mean-square (m.s.) to a random variable A(β) as tt→ 0 iff

4.1. PROCESS OF 2ND ORDER – MEAN-SQUARE CALCULUS 95

2 limEXtAβ ⎡⎤()() ;ββ−= 0 ■ (2) tt→ (⎣⎦) 0

For the sake of convenience, the m.s. convergence will be denoted as

2 limEXtAβ ⎡⎤()() ;ββ−=⇔= 0 l.i.m. Xt() ; ββ A () (3) tt→ (⎣⎦) tt→ 0 0

Remark 1.5 : Necessary and sufficient condition for the m.s.-convergence of the s.p. X ()t ;β , tT∈ , when tt→ 0 , is the Cauchy criterion,

β 2 limEXt⎡⎤()()12 ;ββ−= Xt ; 0 (4) tt→ (⎣⎦) 10 tt20→ since the space of second order stochastic processes is a Hilbert space (Theorem 1.3). ■

Theorem 1.6 : Let X ()t ;β , tI∈ be a second-order stochastic process. A process X ()t ;β converges in the mean-square to a random variable A(β) as tt→ 0 iff the covariance function

RXX ()tt12, convergence to a finite limit as tt10→ , tt20→ .

ms.. Proof : Let Xt();ββ⎯⎯→ A (), as tt→ 0 . Then,

EXtββ;;βββ Xt−= A2 E⎡ Xt ; ββ − A⎤⎡ Xt ; ββ −+ A ⎤ ()()()()12(⎣ ()()()() 1⎦⎣ 2 ⎦)

ββ +−+−EXt()⎣⎦⎣⎦⎡⎤⎡⎤()()()12;;βββ A A EXt()( βββ ) A () A ()

Using Schwartz inequality one obtains

R tt,;;=→ Eββ Xtββ Xt E A β2 as tt→ , tt→ . XX ()12() ( 1 )() 2 ( ()) 1020

Conversely, let RXX ()tt12, → c as tt10→ , tt20→ , where c <∞. Then,

2 EXtβ ⎡⎤;;ββ−= Xt (⎣⎦()()12) =+−EXtββ;;2;;ββ22 EXt EXt β ββ Xt = ()()12()() ()()() 12

=+RttRttXX()11,,2,20 XX ( 2 2 ) − Rttccc XX () 12 →+−=

as tt10→ , tt20→ . This relationship and the Cauchy m.s. criterion imply that ms.. Xt();ββ⎯⎯→ A (), as tt→ 0 . ■

Preposition 1.7 : The most important properties of m.s.-convergence are a) For deterministic functions the m.s.-convergence is equivalent with the simple convergence. b) The operator l.i.m. is linear , i.e. if a and b are deterministic constants, X ()t ;β , Yt();β , tT∈ two second-order stochastic processes and the limits

l.i.m.Xt() ; β = X0 ()β and l.i.m.Yt( ; β ) = Y0 (β ) exists, then tt→ 0 tt→ 0

l.i.m.{aX() t ;ββ+=+ bY( t ; )} aX00( ββ) bY ( ) (5) tt→ 0

96 CHAPTER 4 STOCHASTIC CALCULUS – PRINCIPLES AND RESULTS

β c) If l.i.m.Xt() ; β = X0 ()β exists, then the limit limEXt( ( ; β )) exists and the tt→ 0 tt→ 0 following relation holds,

EXtEXtββl.i.m.() ;β = lim()() ;β (6) ()tt→ 0 tt→ 0

i.e. the operators E β ()i and l.i.m.[i] are permutable. d) If X ()t ;β , Yt();β , tT∈ are two second-order stochastic processes and the limits

l.i.m.Xt() ; β = X0 ()β and l.i.m.Yt( ; β ) = Y0 (β ) exists, then the limit tt→ 0 tt→ 0 limEXtβ ()() ; β exists and the following relation holds, tt→ 0

EXtYtEXtYtββl.i.m.()() ;β ;βββ= lim()()() ; ; (7) ()tt→ 0 tt→ 0 Proof : The above assertions are direct sequences of the definition of m.s.-convergence. For details we refer to ATHANASSOULIS, G.A., (Stochastic Modeling and Forecasting of Ship Systems. Lecture Notes NTUA). ■

4.1.3. Mean-Square Continuity

Definition 1.8 : A second-order stochastic process X ()t ;β ,tI∈ , is continuous in a mean square sense (m.s.-continuous) at tI∈ , iff

β 2 limEXtXt⎡⎤() ;ββ−= (0 ; ) 0 (8) tt→ (⎣⎦) 0

A s.p. X ()t ;β is m.s.-continuous on I , if it is m.s.-continuous at every point tI∈ . ■

Theorem 1.9 : (m.s.-continuity criterion) A second-order s.p. X ()t ;β ,tI∈ is m.s.-continuous at tI∈ , iff the covariance function RXX ()tt12, is continuous at tt12= .

Proof : Let RXX ()tt12, be continuous at tt12= . Then

ββ2 E(⎣⎦⎣⎦⎣⎦⎡⎤⎡⎤⎡⎤ Xt()()+− h;;ββ Xt) = E( Xt()()()() +− h ;; ββ Xt ⋅+− Xt h ;; ββ Xt ) = h→0 =++−+−++⎯⎯⎯RXX()()()() t ht,,,,0 h R XX t ht R XX tt h R XX tt → therefore X ()t ;β is m.s. continuous at t . Let us assume that X ()t ;β is m.s.-continuous at t . Since

β RthtkRttEXX()()++−, XX , =(⎣⎡ Xth( + ;;ββ) − Xt( )⎦⎣⎤⎡ ⋅ Xtk( + ;; ββ)() − Xt ⎦⎤) +

β ++−E()⎣⎦⎣⎦⎡⎤⎡⎤ Xt( h;;;βββ ) Xt ()() Xt ++− Xt ( k ;;; βββ ) Xt ()() Xt

use of the Schwartz inequality implies that RXX ()tt12, is continuous at tT∈ . ■

4.1. PROCESS OF 2ND ORDER – MEAN-SQUARE CALCULUS 97

Theorem 1.10 : If the covariance function RXX ()tt12, is continuous at every diagonal point tt12= then it is continuous everywhere on I ×I .

Proof :

RthtkRttXX()()12++−,, XX 12 =

ββ =+E() Xt()()12 h;;ββ Xt +− k E() Xt()() 12 ;; ββ Xt ≤

β ≤+−⋅++E()⎣⎦⎡⎤ Xt()()()112 h;;ββ Xt Xt k ; β

β ++−⋅E()⎣⎦⎡⎤ Xt()()()221 k;;;βββ Xt Xt

Using the Schwartz inequality it is easily seen that the above expression tends to zero as h → 0 and k → 0 . ■

Remark 1.11 : It should be emphasized that m.s.-continuity of a s.p. X ()t ;β ,tI∈ implies that it is continuous with respect to probability. Indeed the Tchebycheff inequality gives

2 β +−ββ EXth()()()00;; Xt c ()Xt()()00+− h;;ββε Xt >≤ (9) ε2

The m.s.-continuity does not, however, imply continuity of the realizations of the process. As an example one can take the Poisson process. Its covariance function RXX ()tt12,min,= ν () tt 12 is continuous for all tt12,0> , but almost all realizations of this process have discontinuities over a finite interval of time (see Chapter 2/Section 2.2.1.c). ■

4.1.4. Mean-Square Differentiation

Definition 1.12 : A second-order s.p. X ()t ;β ,tI∈ , has mean-square derivative X ′()t ;β at tI0 ∈ , if there is a r.v. Xt′()0 ;β such that

2 ⎛⎞⎡⎤Xt+− h;;ββ Xt β ⎜⎟()()00 limEXt⎢⎥= ′()0 ; β (10) h→0 ⎜⎟h ⎝⎠⎣⎦

A s.p. X ()t ;β is m.s.-differentiable on I , if it is m.s.-continuous at every point tI∈ . Higher order derivatives can be defined analogously. ■

Theorem 1.13 [m.s.-Differentiation criterion] : A second-order s.p. X ()t ;β ,tI∈ is m.s.- differentiable at tI0 ∈ , iff there exist the second-order derivative of the covariance function ∂2 RXX ()tt12, . ∂∂tt12

Proof : This theorem follows directly from Theorem 1.6, which asserts that condition (10) is satisfied iff

98 CHAPTER 4 STOCHASTIC CALCULUS – PRINCIPLES AND RESULTS

β E ()Yhk⋅⎯⎯⎯ Yk→0→ finite limit, h→0 where,

1 YXthXt=+−⎡⎤()();;ββ (11) h h ⎣⎦

But

ββ⎛⎞1 EYY()hk⋅= E⎜⎟⎡⎤⎡⎤ Xth()()()() +;;ββ − Xt ⋅ Xtk + ;; ββ − Xt = kh ⎣⎦⎣⎦ ⎝⎠ 1 =++−+−++()RXX(thtkR,, )()()() XX ttkR XX thtR ,, XX tt kh

Therefore, in order to ensure the existence if the m.s.-derivative X ′(t ;β) of the s.p. X ()t ;β is necessary and sufficient that the above expression has finite limit, i.e. that the second-order derivative defined as

∂2 ⎡⎤1 RXX()tt12,lim=++−+−++⎢⎥() RthtkRttkRthtRtt XX()()()() , XX , XX , XX , ∂∂ttk→0 ⎢⎥ kh 12 h→0 ⎣⎦ exists and is finite at the point ()tt, . ■

Preposition 1.14 : The m.s.-derivatives have the following properties:

a) m.s.-differentiability of X ()t ;β at tI0 ∈ , implies m.s.-continuity of X ()t ;β at tI0 ∈ . b) If X ()t ;β , Yt( ;β) , tI∈ are two second-order s.p. m.s.-differentiable, then the m.s.- derivative of aX() t;;β + bY () t β exists and

d dX( t;;β ) dY( t β ) {}aX() t;;ββ+= bY () t a + b (12) dt dt dt

where a and b are constants. c) If X ()t ;β , tI∈ is n-times m.s.-differentiable on I , the means of these m.s.- derivatives of X ()t ;β exist on I and moreover the following equation holds,

⎛⎞dXtn ( ;β ) d n EEXtββ= ;β (13) ⎜⎟nn()() ⎝⎠dt dt

Proof : a) is direct consequence of Theorems 1.9 and 1.13. b) comes as result of Preposition 1.7b and the definition of m.s.-derivative. c) can be proved using Preposition 1.7c and the definition of m.s.-derivative. ■

Theorem 1.15 : If X ()t ;β ,tI∈ has m.s.-derivative for all tI∈ , then the partial derivatives ∂∂∂R ()tt,, R() tt2 R() tt , XX12,, XX 12 XX 12 exist and ∂∂∂∂tttt1212 ∂R ()tt, β ′ XX 12 RttEXtXtXX′ ()12,;;==() ()() 1ββ 2 (14) ∂t1 4.1. PROCESS OF 2ND ORDER – MEAN-SQUARE CALCULUS 99

∂R ()tt, β ′ XX 12 RttEXtXtXX ′ ()12,;;==() ()() 1ββ 2 (15) ∂t2 ∂R ()tt, β ′′ XX 12 RXX′′() tt12,;;== E() Xt ()() 1ββ Xt 2 (16) ∂∂tt12

Proof : If X ()t ;β ,tI∈ has m.s.-derivative X ′(t ;β) at tI∈ , then 1 ()X ()()th+−;;ββ Xt ⎯⎯→ms.. Xt′() ; β as h → 0 h Therefore, RthtRtt()()+−,,⎛⎞ Xth()() +− ;;ββ Xt XX12 XX 12β ⎜ 1 1 ⎟ lim== limEXt⎜ ()2 ;β ⎟ hh→→00hh⎝⎠⎜ ⎟ ⎛⎞Xt()()+− h;;ββ Xt ββ⎜ 11⎟ ′ ==EXtEXtXt⎜l.i.m.()212 ;βββ⎟ () ()() ; ; ⎝⎠⎜ h→0 h ⎟ ∂R ()tt, The above implies that the derivative XX 12 exists and eq.(14) holds. ∂t1 ∂R ()tt, Similarly, it can be shown that the derivative XX 12 exists and the formula (15) holds. ∂t2 Finally,

1 ⎛⎞⎛⎞∂+∂Rttk(),, Rtt() Xth()() +− ;;ββ Xt ⎜⎜XX12 XX 12⎟⎟β ′ 2 2 lim⎜⎜−=⎟⎟ limEXt()1 ;β = kk→→00⎜⎜⎟⎟ kt⎝⎠⎝⎠⎜⎜∂∂11 t k β = EXt()′′()()12;;ββ Xt shows that formula (16) is valid. ■ Remark 1.16 : Analogous results for the derivatives of higher order can be easily derived. The covariance function of the derivatives of order k and l of a given process is expressed by the general formulas

kl+ β ()kl() ∂ R ()kl ()()tt,;;,== E X() tββ X() t R() tt (17) XX 12() 1 2klXX 12 ∂∂tt12 if appropriate m.s.-derivatives exist. ■

Let us assume that RXX ()tt12, has an infinite number of derivatives on I×I . Let ttn Xt();0;0;ββ=+ X ( ) X′() β ++… X()n () 0; β. Simple transformation of the expression n 1!n 2 β ββ− EXt()();; Xtn () leads to the following

Theorem 1.17 : A second-order s.p. X ()t ;β ,tI∈ is m.s.-analytic iff its covariance function

RXX ()tt12, is analytic at each diagonal point ()tt, ∈× I I. ■

Remark 1.18 : As in the case of m.s.-continuity it should be emphasized that m.s.- differentiation of a s.p. X ()t ;β ,tI∈ does not implies differentiation of the realizations of the 100 CHAPTER 4 STOCHASTIC CALCULUS – PRINCIPLES AND RESULTS

stochastic process. For the conditions required for differentiation of the realizations see Section 2. ■

Remark 1.19 : It follows directly from Theorem 1.13 that, a weakly-stationary s.p.

X ()t ;β ,tI∈ is differentiable at tI0 ∈ iff its covariance function RXX ()τ , τ =−tt12, has the derivative of a second order at τ = 0 . In this case formula (17) reduce to the following one kl+ β ()kl() k d RttEXtXt()kl ()()12,;;1==−()() 1ββ() 2 () kl+ RXX ()τ ■ (18) XX dτ

4.1.5. M.S. Integration I – Integration over an Independent Stochastic Process

Let the second-order stochastic processes X ()t ;β , tI∈ and Ψβ()t ; , tI∈ . In what follows we will consider, for various cases of the above s.p.’s, the meaning of the integral

I ()β = ∫ Xt ()();;βΨ d t β (19) I One very important case, that gives us many of the known integrals, of practical importance is when the s.p.’s X ()t ;β , tI∈ and Ψβ()t ; , tI∈ , are mutually independent, i.e.

EXtβββ()()();;ββ⋅=Ψ tEXtE()() ; β ⋅()Ψ ()t ; β

For this case we have the following

Theorem 1.20 : Let X ()t ;β , tI∈ and Ψβ()t ; , tI∈ be orthogonal second-order stochastic processes. Then the s.p. X ()t ;β is integrable with respect to Ψβ()t ; iff

1. RXX ()tt12, is integrable on I ×I

2. RΨΨ ()tt12, has bounded variation on I ×I In this case we have,

2 ⎛⎞⎡⎤⎟ β ⎜ ⎟ 2 E⎜⎢⎥ Xt();;ββ dΨ () t⎟= R ( tt , ) dR ( tt , ) (20) ⎜⎢⎥∫∫∫⎟ XX 12ΨΨ 12 ⎝⎠⎜⎣⎦⎢⎥III⎟ ×

Proof : Let []ab, ⊂ I be a finite interval and let the points tt01,,… , tn define a partition of [ab, ] such that

at=<<<=01 t… tn b

⎡⎤ Let ξiii∈ ⎣⎦tt, +1 . We form the random variable

n−1 I βξβΨβΨβ=−Xt;;;⎡⎤ t(21) niii()∑ ( )⎣⎦⎢⎥()+1 () i=1

2 By virtue of Theorem 1.6 IIββ⎯⎯→ms.. iff EEββIIββ⎯⎯→ms.. ⎡⎤ I β. Thus n () () ()nm() () ()⎣⎦() we have from eq. (21),

4.1. PROCESS OF 2ND ORDER – MEAN-SQUARE CALCULUS 101

β E ()IInm()ββ ()= mn−−11 (22) =−−EXβ ξβnmm;; X ξβΨ⎡ t ; β Ψ t m ; βΨ⎤⎡ t n ; β Ψ t n ; β ⎤ ∑∑ ()()()()()()()ijj⎣⎢ ++11 ji⎦⎣⎥⎢ i ⎦⎥ ji==11

Since X ()t ;β and Ψβ()t ; are orthogonal we will have

β E ()IInm()ββ ()= mn−−11 (23) =−−+R ξξnm,,⎡⎤Rt mn t Rt mn ,, t Rtt mn Rtt mn , ∑∑ XXi()( j⎣⎦⎢⎥ΨΨ j++11 i ) ΨΨ () j + 1 i ΨΨ () ji + 1 ΨΨ () ji ji==11

It can be shown that the above sum converges iff the conditions mentioned above holds. ■

Some basic properties of the integral of a s.p. over an independent s.p. are as follows,

Theorem 1.21 : Let X i (t ;β), tI∈ and Ψβi ()t ; , tI∈ be orthogonal second-order stochastic processes with the assumptions of Theorem 1.20. Then

a) aXtaXtdtaXtdtaXtdt;;;;;;;βββββββ+=ΨΨΨ + ∫∫∫()11() 21 () () 1 1 () () 2 2 () () III

b) X tdat;;ββΨΨ+= a t ; β aXtdt ;;;; ββββ Ψ + aXtdt Ψ ∫∫∫()()11 () 2 2 () 1() 1 () 2 () 2 () III

⎛⎞ β ⎜ ⎟ c) EXtdt⎜ ();;ββΨ ()⎟= mtdmtX ()()Ψ ⎜∫∫⎟ ⎝⎠II

d) Moreover if X ()t ;β is differentiable with derivative X ′(t ;β), the following integration by parts formula holds

b X ()tdt;βΨΨ () ; β=− Xt ()() ;; ββ t Xt′()() ;; ββ Ψ tdt ∫∫a II

Proof : Assertions a), b) and c) follow directly, by considering the sequence of eq. (21). For the proof of d) we refer to SOBCZYK, K., (Stochastic Differential Equations. p.111) ■

We will now study specifc integrals that come from different cases of orthogonal s.p.’s X ()t ;β , tI∈ and Ψβ()t ; , tI∈ .

I. Mean Square Integral of the s.p. X ()t ;β with respect to the function ψ ∈ CI0 (),

By settingΨβ(tt; )≡ ψ( ), tI∈ where ψ ∈ CI0 (), the integral (19) becomes

I ()ββψ= ∫ X (tdt; ) () (24) I

By virtue of Theorem 1.20 the above integral exists iff

• RXX ()tt12, is integrable on I ×I • ψ()t has bounded variation on I

In this case eq. (20) takes the form, 102 CHAPTER 4 STOCHASTIC CALCULUS – PRINCIPLES AND RESULTS

2 ⎛⎞⎡⎤⎟ β ⎜ ⎟ E⎜⎢⎥ Xtdt()();,βψ⎟= R ( ttdtdt )()() ψ ψ (25) ⎜⎢⎥∫∫∫⎟ XX 12 1 2 ⎝⎠⎜⎣⎦⎢⎥III⎟ ×

II. Riemann-Stieltjes Integral of the function fCI∈ 0 (), with respect to the s.p. Ψ (t;β )

By setting X ()tft;β ≡ (), tI∈ where fCI∈ 0 (), the integral (19) becomes

I ()β = ∫ f ()tdΨβ ( t; ) (26) I

By virtue of Theorem 1.20 the above integral exists iff

• f ()t is integrable on I

• RΨΨ ()tt12, has bounded variation on I ×I

In this case eq. (20) takes the form,

2 ⎛⎞⎡⎤⎟ β ⎜ ⎟ 2 Eftdt⎜⎢⎥()Ψ (;,β )⎟= ftftdRtt ( ) ( ) ( ) (27) ⎜⎢⎥∫∫∫⎟ 12ΨΨ 12 ⎜⎝⎠⎣⎦⎢⎥III⎟ ×

III. Riemann-Stieltjes Integral of the s.p. X ()t ;β with respect to the martingale Ψβ()t ; (Orthogonal Increment Process).

2 We will study the form of the quantity Δ hw,12RΨΨ (tt, ) as hw,0→ . We have

Δ2 RttE,;;;;=+−⋅+−β ⎡⎤⎡⎤ Ψ th βΨ t βΨ t w βΨ t β hw,12ΨΨ ()()⎣⎦⎣⎦ ( 1 )()( 1 2 )( 2 )

For every tt12≠ we can find appropriately small hw , , such that

Ethβ ⎡⎤⎡⎤ΨβΨβΨβΨβ+−;; t ⋅ tw + ;;0 − t = ()⎣⎦⎣⎦(1122 )()( )()

On the other hand for tt12= we will have

ΔΨβΨβΨβΨβ2 =+−⋅+−=β ⎡⎤⎡⎤ hw, RttEΨΨ (),;;;;()⎣⎦⎣⎦ ( th ) () t ( tw ) () t 2 =+β ⎡⎤ΨβΨβ − = Et()⎣⎦⎢⎥()min() hwt , ;() ;

=+Ft()min() hw , − Ft() =Δmin()hw , Ft() where the function is defined at Chapter 2/Section 2.2.1.c. So, the conditions for the existence of the integral are fulfilled iff

• RXX ()tt, is integrable on I • F (t) has bounded variation on I

In this case eq. (20) takes the form,

4.1. PROCESS OF 2ND ORDER – MEAN-SQUARE CALCULUS 103

2 ⎛⎞⎡⎤⎟ β ⎜ ⎟ EXtdt⎜⎢⎥();;ββΨ ()⎟= RttdFt ()() , (28) ⎜⎢⎥∫∫⎟ XX ⎝⎠⎜⎣⎦⎢⎥II⎟

4.1.6. M.S. Integration II–Integration over Martingales with non-anticipative dependence

Another important case of stochastic integral is when the stochastic process is being integrated over martingales. One special case with various applications in Mechanics is the Ito Integral which, as we shall see, is the integration of a s.p. over a Wiener Measure. The essential feature of the construction of this integral is that the dependence of the stochastic process on the martingale should be non-anticipative; the stochastic process can depend at most, on the present and past values of the martingale process.

Let Ψβ()t ; be a second order, martingale, U t()Ω - measurable. For every ts, ∈ I we define the stochastic process ηβΨβΨβ()()()ts,;=− t ; s ; . Using Prepositions 2.39-2.40/Chapter 2, we clearly have

a) Etsβ ()ηβ(),;= 0, for every ts , ∈ I β ⎡⎤2 b) EtsFtFs()⎣⎦ηβ(),; =−<∞() () , for every st< with ts, ∈ I β c) Ets()ηβ()(),; |U τ Ω= 0, for every ts , ,τ ∈ I with τ < min{}ts , .

Let as denote by HI2 () the class of stochastic processes X ()t ;β ,tI∈ satisfying the following conditions,

a) X ()t ;β is U t()Ω - measurable, ⎡⎤2 b) the integral ∫ ⎣⎦X ()tdt;β is finite with probability one. I Remark 1.22 : The last two conditions and the Preposition 2.40d/Chapter 2 , implies that for every τ ∈ I the process X (τβ; ) is independent of the increments ηβ()ts,; for all ts , ∈ I such that τ < min{}ts , .

Theorem 1.23 : Let X ()ii; ∈ HI2 (). Then there is a sequence of simple random functions X n ()ii; ∈ HI2 () such that

⎡⎤ 2 limXtn () ;ββ−= Xt () ; dt 0 (29) n→∞ ∫ ⎣⎦⎢⎥ I

Proof : We will now introduce the stochastic integral over a martingale process in two stages. First we define it for a simple function. Let X ()ii; ∈ HI2 () be a simple function, i.e. Xt;;ββ= Xt for ttt∈ , ⎤ where t n is a partition of I . In this case the above () (i ) ( ii+1⎦ {}i i=1 integral is defined as follows:

n−1 Xt;;βΨ d t β=− Xt ; βΨ⎡ t ; β Ψ t ; β⎤ (30) ∫ ()()∑ (ii )⎣⎢ ()+1 () i⎦⎥ I i=1 104 CHAPTER 4 STOCHASTIC CALCULUS – PRINCIPLES AND RESULTS

Since for every function X ()ii; ∈ HI2 () there is a convergent sequence of simple functions X n ()ii; ∈ HI2 () we form the sequence n−1 I ==Xt;;βΨ d t β Xt ; βΨ⎡ t ; β − Ψ t ; β⎤ (31) nn∫ ()()∑ nii ( )⎣⎢ ()+1 () i⎦⎥ I i=1

Definition 1.24 : Let the stochastic process X ()ii; ∈ HI2 () and X n ()ii; ∈ HI2 () a sequence of simple functions. Let also In the sequence of random variables of eq. (31). If

l.i.m. InM= I (32) n→∞ Δ→n 0

exists, the random variable IM is called the integral of the stochastic process X ()ii; ∈ HI2 () over the martingale Ψβ()t ; . ■

Remark 1.25 : We note that, for the above integral we use the values of the stochastic process at ti . ). The value of the m.s. Integral of a stochastic process over a martingale process depends on the choice of the intermediate points ξi . To illustrate this dependence let consider the two integrals on the interval I =[]ab,

n−1 ⎡⎤ IMiii,0=−l.i.m.ΨβΨβΨβ()tt ;()+ 1 ;() t ; n→∞ ∑ ⎣⎦⎢⎥ Δ→n 0 i=1 n−1 ⎡⎤ IMiii,1=−l.i.m.ΨβΨβΨβ()()tt++ 1 ; 1 ;() t ; n→∞ ∑ ⎣⎦⎢⎥ Δ→n 0 i=1

Getting the difference of the two above integrals, we have

n−1 ⎡⎤2 Levy Property IIMM,1−= ,0l.i.m.ΨβΨβ()tt i+ 1 ; −() i ; ⎯⎯⎯⎯⎯→− FbFa ()() ■ n→∞ ∑ ⎣⎦⎢⎥ Δ→n 0 i=1 Remark 1.26 : An alternative approach is also been used by some authors based on the convergence of the above limits with respect to probability instead of mean-square calculus. (SOIZE, C., The Fokker-Planck Equation and its Explicit Steady State Solutions). ■

Theorem 1.27 : Let X ()ii; ∈ HI2 (), I =[]ab, and Ψβ()t ; a second-order, U t()Ω - measurable, martingale with continuous characteristic F (t). Then the integral IM as defined above exists and is unique, in the mean-square sense. Moreover

2 ⎛⎞⎡⎤⎟ β ⎜ ⎟ EXtdt⎜⎢⎥();;ββΨ ()⎟= RttdFt ()() , ⎜⎢⎥∫∫⎟ XX ⎝⎠⎜⎣⎦⎢⎥II⎟

ms.. Proof : By virtue of Theorem 1.6 IIn ()ββ⎯⎯→ () iff 2 ββββ⎯⎯→ms.. ⎡⎤ β EE()IInm() () ()⎣⎦ I() . Thus we have from eq.(31),

β E ()IInm()ββ ()= mn−−11 (33) =−−EXtβ nm;;ββΨβΨβΨβΨβ Xt⎡ t m ; t m ;⎤⎡ t n ; t n ; ⎤ ∑∑ ()()()()()()()ij⎣⎢ j++11 j⎦⎣⎥⎢ i i ⎦⎥ ji==11 4.1. PROCESS OF 2ND ORDER – MEAN-SQUARE CALCULUS 105

It is to be evaluated over the region [][]ab,,× ab as indicated in the figure below. Each term in the sum is to be evaluated over a rectangle such as R1 or R2 . In evaluating these terms, we distinguish two types of rectangles: those containing a segment of the diagonal line ttnm= , such as R1 and those containing at most one point of the diagonal line, such as R2 .

t n

n t i+1 R 2 n t i R1

m m m t j t j+1 t

mm nn For the second type, a typical situation is one where ttj <≤

EXtβ nm;;ββΨβΨβΨβΨβ Xt⎡⎤⎡⎤ t m ;−−= t m ; t n ; t n ;0 ()()()()()()()ij⎣⎦⎣⎦⎢⎥⎢⎥ j++11 j i i

Hence, the contributions of all the terms in eq. (33) over rectangles of the second type are zero. Turning now to rectangles of the first type, consider a typical situation where mnm n tttj <

EXtβ nm;;ββΨβΨβΨβΨβ Xt⎡⎤⎡⎤ t m ;−−= t m ; t n ; t n ; ()()()()()()()ij⎣⎦⎣⎦⎢⎥⎢⎥ j++11 j i i 2 =−=EXtβ nm;;ββΨβΨβ Xt⎡⎤ t m ; t n ; ()()()()()ij⎣⎦⎢⎥ j+1 i =⋅−EXtβ nm;;ββ Xt⎡⎤ Ft m Ft n ()()()ij⎣⎢ () ji+1 ()⎦⎥

Thus, mn−−11 EββIIββ=⋅− E Xtnm;; β Xt β⎡ Ft m Ft n⎤ ()nm() () ∑∑ ()()() ij⎣⎢ () ji+1 ()⎦⎥ ji==11 So, we have the same summation as in case III of the above paragraph. Since F (t) is continuous, it has finite variation on I . Moreover X ()ii; ∈ HI2 (), so RXX (tt, ) is integrable on I . ■

Some basic properties of the integral of a s.p. over a martingale are as follows,

Theorem 1.28 : Let X i ()ii; ∈ HI2 () and Ψβi ()t ; , tI∈ , I =[]ab, , is a second-order, U t()Ω - measurable, martingale with continuous characteristic F (t). Then

106 CHAPTER 4 STOCHASTIC CALCULUS – PRINCIPLES AND RESULTS

a) ∫ Xt();;0ββ dΨ () t = I b) aXtaXtdtaXtdtaXtdt;;;;;;;βββββββ+=ΨΨΨ + ∫∫∫()11() 21 () () 1 1 () () 2 2 () () III c) X tdat;;ββΨΨ+= a t ; β aXtdt ;;;; ββββ Ψ + aXtdt Ψ ∫∫∫()()11 () 2 2 () 1() 1 () 2 () 2 () III b d) X ()tdt;βΨΨ () ; β=− Xt ()() ;; ββ t Xt′()() ;; ββ Ψ tdt if X ()ii; is m.s. ∫∫a II differentiable. τ e) χΨβΨτβΨαβtd t;;;=− ∫ []ab, () ( ) ( ) ( ) α

f) The integral IM of X i (t ;β) over Ψβ()t ; is also a martingale with characteristic τ GRttdFtτ = , ()∫ XX ( ) () α τ g) The sample functions of the s.p. I τβ;;;= Xt β d Ψ t β are continuous with M ()∫ ()() α probability one.

Proof : Assertions a), b) and c) follow directly, by considering eq. (31). The proof of d) and e) can be found at SOBCZYK, K., (Stochastic Differential Equations. p.111 and 108 respectively). Proof of f) and g) can be found at GIΚHMAN, I.I. & SKOROΚHOD, A.V., (The Theory of Stochastic Processes III. p. 54 and 57 respectively.). An alternative proof of the last assertion could be based on direct application of the Kolmogorov Theorem with the use of mean value theorem for integrals. ■

Remark 1.29 : As we have mention at the beginning a very important case of stochastic integral, for applications, is the Ito Integral. For this case, the martingale is a standard Wiener process U t()Ω -measurable. The Ito Integral can be defined for every process X ()ii; ∈ HI2 (), since the characteristic of the standard Wiener process Ft()=∈ t, t I, has bounded variation on I . Of course all the properties derived above for the general case of a martingale process, holds for the Ito Integral. ■

4.1.6.a. A formula for changing variables (Ito Formula)

In this section analogs and corollaries of the chain rule of differentiation of composite functions are studied in the case when differentiation is interpreted as the inverse operation of stochastic integration. The results obtained are of importance in what follows.

Theorem 1.30 [Ito Formula]: Let

1. ψβ0 () be a second-order random variable, 2. at();β , tI∈ be a second-order stochastic process that its correlation function

Rαα ()tt12, has bounded variation on I ×I , 4.1. PROCESS OF 2ND ORDER – MEAN-SQUARE CALCULUS 107

3. μβ()t ; , tI∈ be a second-order, martingale, U t()Ω - measurable with continuous characteristic Ft( ), tI∈

4. Ψβ()tatt;;;=+ ψβ0 ()() βμβ + (), tI∈

We also assume that fx( ), x∈ is a twice continuously differentiable function. Then

df1 d2 f f Ψβtf;;;;=+ ψβ ΨβΨβ sds + Ψβ sdFs (34) ()() ()0 () ∫∫()() () 2 ()()() IIdx2 dx

Proof : For the proof we refer to GIΚHMAN, I.I. & SKOROΚHOD, A.V., (The Theory of Stochastic Processes III. p. 70). ■

Corollary 1.31 : If f ()xt,, x∈∈ , t I, is continuously differentiable with respect to tI∈ and twice continuously differentiable with respect to x ∈ , then

f ()Ψβ()ttf;,−=+() ψβ0 () , tatbt() ; β () ; β (35) where

∂∂ff1 ∂2 f a() t;;,;,;;,βΨβΨββ=+()() s t ds()() s s da() s +() Ψβ() s s dF() s ∫∫∂∂tx2 ∫ ∂ x2 II I and ∂f bt();;,;βΨβμβ= ()() s sd() s ∫ ∂ I x

Proof : For the proof we refer to GIΚHMAN, I.I. & SKOROΚHOD, A.V., (The Theory of Stochastic Processes III. p. 71). ■

Concerning the special case of Ito Integral, i.e. integration over a standard Wiener process we have the following result

Corollary 1.32 : Let Wt();,β t∈ I be a standard Wiener process, U t()Ω -measurable and fxx(), ∈ is a twice continuously differentiable function. Then df1 d2 f f ()Wt();0ββββ=+ f() () Ws() ; dWs() ; + () Ws() ; ds (36) ∫∫dx2 dx2 II

Proof : Direct application of Theorem 1.30. ■

4.1.6.b Stochastic Differentials

Definition 1.33 : Let 1. at();β , tI∈=[ ab, ] be a second-order stochastic process that its correlation function

Rαα ()tt12, has bounded variation on I ×I ,

2. μβ()t ; , tI∈ be a second-order, martingale, U t()Ω - measurable with continuous characteristic Ft( ), tI∈ . 108 CHAPTER 4 STOCHASTIC CALCULUS – PRINCIPLES AND RESULTS

We say that a stochastic process X ()t ;β , tI∈ possesses a stochastic differential (of a continuous type)

dXt()(;,βϕ=+ dad )( ψμϕβ , ) = ()()()() tdat ;; βψβμβ + tdt ;; , tI∈ (37)

where, ψ()ii; ∈ HI2 () and the process ϕβ()t ; is progressively measurable with respect to

U t()Ω whose realizations are (with probability one) bounded functions iff tt X ()tX;0;;;ββϕββψβμβ=+ ()∫∫ ()() sdassds + ()() ;; (38) aa for each and every tI∈ . ■

Remark 1.34 : We recall that a stochastic process ϕβ()t ; , tI∈=[ ab, ] is progressively measurable iff the restriction of the function ϕβ()t ; to the set [as, ]×Ω is UUss()I ×Ω ( ), i.e. if

{}()()ttBtasI,:βϕ ; β∈∈ , [] , ∈UUss() × ( Ω ) (39) for every B ∈U ()DT() and sT∈ . ■

Clearly a process possessing a stochastic differential of continuous type has continuous modification. In what follows we shall consider just this kind of modification of the process Ψβ()ttI;,∈ . Theorem 1.30 can be formulated using the concept of a stochastic differential in the following manner. If process Ψβ()t ; , tI∈ possesses stochastic differential dtΨβ();;;=+ dat () β dt μβ () and the function f ()xt,, x∈∈ , t I, is continuously differentiable with respect to tI∈ and twice continuously differentiable with respect to x ∈ then the process X ()tftttI;;,,βΨβ=∈() () also posses a stochastic differential and, moreover, for every tI∈ ,

∂∂ff dX() t;;,;,;βΨβ=+()() t t dt() Ψββ() t t da() t + ∂∂tx 1 ∂∂2 ff ++=()Ψβ()ttdFt;,() () Ψβ() ttdt ;, μβ() ; (40) 2 ∂∂xx2 ∂∂ff1 ∂2 f =+()Ψβ()ttdt;,() Ψβ() ttdt ;, Ψβ() ; +() Ψβ() ttdFt ;, () ∂∂tx2 ∂ x2

This implies

Theorem 1.35 : If the stochastic processes Ψβi (t ;,) tI∈ (we write Ψi for simplicity) possess stochastic differentials dtΨβiii();;;=+ dat () β dt μβ () then

ddd()ΨΨ12+= Ψ 1 + Ψ 2 (41)

ddddFt()ΨΨ12⋅= ΨΨ 1 2 + Ψ 21 Ψ + 12 () (42) ⎛⎞ΨΨΨ−ΨΨddΨΨdF() t− dF () t d ⎜ 12112⎟=+12 212 (43) ⎜ ⎟ 23 ⎝⎠ΨΨ22 Ψ 2 4.1. PROCESS OF 2ND ORDER – MEAN-SQUARE CALCULUS 109

1 de()ΨΨuu=+ e udΨμ e Ψ u2 dFt() (44) 2

where Ft12 () is the joint characteristic of μβ1 (t ; ) and μβ2 ()t ; , i.e.

1 Ft()=−−⎡⎤ F() t Ft () Ft () 122 ⎣⎦⎢⎥S1,2 1 2 ⎛⎞Ψ with Ft() denoting the characteristic of the sum μβμβ()tt;;+ ().The formula for d ⎜ 1 ⎟ S1,2 12 ⎜ ⎟ ⎝⎠Ψ2 is applicable provided Ψβδ2 ()t ;0≥>.

Proof : The formulas follow directly from Theorem 1.30 when applied successively to

x1 xu functions xxxx1+⋅ 212,, and e . ■ x2 110 CHAPTER 4 STOCHASTIC CALCULUS – PRINCIPLES AND RESULTS

4.2. Analytical Properties of Sample Functions

Second order properties of a stochastic process formulated by use of m.s.-convergence in terms of correlation functions do not provide, in general, sufficient information about the behavior of the realizations of a process. This fact can be illustrated by a Poisson process which – as we have seen in Section 2.2.1.c/Chapter 2 – is m.s.-continuous, but its sample functions are discontinuous. It is therefore of interest to throw some light on the analysis and the results associated with sample function properties. As we mention at the beginning of the chapter we assume the existence of a probability space (Ω, Uc(Ω), ) and the stochastic process

X :Ω→DT() defined on it, with induced probability space ()DT(),,Uc() DT () DT() .

4.2.1. Sample Function Integration

In the theory and applications we often deal with ordinary Lebesque Integral of realization of stochastic process. We, now, recall the definition of a measurable stochastic process. Definition 2.1 : A stochastic process X ()t ;β , tI∈ is a measurable stochastic process iff is measurable with respect to the product measure ccDI()× I , where cI is the Lebesque measure on I . ■ The mathematical base for sample function integration gives a theorem following directly from the known Fubini Theorem. Theorem 2.2 : Let X ()t ;β , tI∈ be a measurable stochastic process. Then almost all realizations of X ()t ;β are measurable with respect to the Lebesque measure cI on I . If β EXt()();β exists for tI∈ , then this average defines a cI - measurable function of tI∈ . If A is a measurable set on I and

∫ EXtβ ()();β dt<∞ (1) A then almost all realizations of X ()t ;β are Lebesque integrable on A and

⎛⎞ ββ⎜ ⎟ EXtdtEXtdt⎜ ();;ββ⎟= () () (2) ⎜∫∫⎟ ⎝⎠AA

Proof : According to the assumption X ()t ;β , tI∈ is a measurable stochastic process. So, the Fubini Theorem implies that the ‘intersection’ X ()t ;β is, for almost all β , a measurable function of tI∈ . The Fubini Theorem also implies that if EXtβ ()();β exists, then it is a measurable function of t. Since the average of X ()t ;β is the integral

Eβ ()Xt();;βββ= ∫ Xt()()c d Ω 4.2. ANALYTICAL PROPERTIES OF SAMPLE FUNCTIONS 111

the assumption (1) asserts that the double integral (first – with respect to β , and then with respect to t) of X ()t ;β is finite. This double integral taken in the opposite order is also finite; therefore the integral

∫ X ()tdt;β A is finite for almost all β ∈Ω. This means that almost all realizations of X ()t ;β are Lebesque integrable on I . The Fubini theorem implies that the value of a double integral is independent of the order of the integration; so we obtain (2). ■ Let us note that the basics assumption of the above theorem requires the process X ()t ;β to be measurable. The following theorem gives a condition for an arbitrary stochastic process X ()t ;β to be equivalent to measurable process X ()t ;β .

Theorem 2.3 : If for almost all t (with respect to the measure cI ) a stochastic process X ()t ;β is continuous in probability then a measurable and separable stochastic process X ()t ;β exists which is stochastically equivalent to X ()t ;β .

Proof : See DOOB, J.L., 1953, Stochastic Processes. John Wiley &Sons. ■

Remark 2.4 : The condition of this theorem (continuity of X ()t ;β in probability) is not very restrictive. It is satisfied, for instance, by all processes whose correlation function is continuous. ■

Remark 2.5 : If f (t) is a non-random Lebesque measurable function defined on I with values in and X ()t ;β is a measurable stochastic process, then the sample function integral ∫ f ()tXt (;β ) dt (3) I exists and by virtue of Fubini theorem ⎛⎞⎛⎞ ββ⎜⎜⎟⎟ E⎜⎜ f() t X ( t;;ββ ) dt f ( s ) X ( s ) ds⎟⎟== E f () t f () s X ( s ;; ββ ) X ( t ) dtds ⎜⎜∫∫⎟⎟ ∫∫ ⎝⎠⎝⎠II II (4) = ftfsR tsdtds, ∫∫ () ()XX ( ) II

4.2.2. Sample Function Continuity

We shall now study the conditions that a s.p. should impose, so that its realizations are continuous. The following theorem gives a sufficient condition for a s.p. to have continious realizations. Theorem 2.6 : Let X ()t ;β , tI∈ be a separable stochastic process. If for all tt, + h in the interval I

c ()X ()()th+−;;ββ Xt ≥ gh() ≤ qh() (5) 112 CHAPTER 4 STOCHASTIC CALCULUS – PRINCIPLES AND RESULTS

where gh() and qh( ) are even functions of h , non-increasing as h → 0 and such that ∞ ∞ ∑ g()2−n <∞ and ∑ 22nnq()− <∞ (6) n=1 n=1 then almost all realizations of a process X ()t ;β are continuous on I . Proof : See LOÈVE, M., (Probability Theory II). ■

The above theorem and the Tchebycheff Inequality:

n EXβ ()()β c ()Xk()β ≥≤ k n imply the following corollary which constitutes a useful sufficient condition for the validity of the theorem.

Corollary 2.7 [Kolmogorov Theorem] : Let X ()t ;β , tI∈ be a separable stochastic process such that for certain number p > 0 and for all tt , + h in the interval I p EXthβ ()()()+−;;ββρ Xt ≤() h (7) where a) ρ()hch=>>1+r ,0,0 c r ch b) ρ()hrp=>1+r , log h then almost all realizations of a process X ()t ;β are continuous on I . Proof : After application of the Tchebycheff Inequality, condition (5) will be satisfied if

ρ()h qh()= ⎡⎤p ⎣⎦gh() r 1+−rap Taking in case a) gh()=<< ha ,0 a one obtains qh()= ch . Taking the case b) p −β r h gh()=<

Example 2.8 : Simple calculations show that for the Wiener process z 11⎪⎧ ξ 2 ⎪⎫ WDIWtWsxi ∈−<=−⋅: exp dξ cDI()(){}() ( ) () ( ) ∫ ⎨ ⎬ −∞ 2π ts− ⎪⎩⎭2 ts− ⎪ and +∞ 42⎧⎫ 4 ξξ⎪⎪1 2 EXthβ ()()+−;;ββ Xt = exp3⎨⎬⎪⎪ −⋅ d ξ =−() ts ()∫ ⎪⎪2 ts− −∞ 2π ts− ⎩⎭⎪⎪ which means that,

4.2. ANALYTICAL PROPERTIES OF SAMPLE FUNCTIONS 113

4 Eβ ()Xt()()+− h;;3ββ Xt = h2

Therefore Kolmogorov condition a) is satisfied with constants c = 3, p = 4, r =1. It proves that almost all realizations of the Wiener process are continuous on each finite interval. ■

If sample functions of a s.p. are discontinuous it is of interest to know whether the discontinuities are of the first kind only (finite jumps). The following theorem holds.

Theorem 2.9 [Chentsov] : Let X ()t ;β , tI∈ be a separable stochastic process. If for ttt123<< with ttt123,,∈ I and tt31−= h

p β ⎡⎤⎡⎤ββββ−−≤1+r EXt()⎣⎦⎣⎦()()()()3221;;;; Xt Xt Xt ch (8) where p,,rc are positive constants, then almost all realizations of X ()t ;β do not have discontinuities the second kind.

Proof : CRAMER H. , LEADBETTER M. R., 1953, Stationary and Related Stochastic Process, Wiley NY.

Remark 2.10 : The conditions of the above theorem are, in particular, satisfied for a process with independent increments and such that for any tt , +∈ h I,

2 Eβ ()Xt()()+− h;;ββ Xt ≤ Ah (9) where A is a constant. It can be verified that Poisson process satisfies the condition.

4.2.3. Sample Function Differentiation

We shall now state here a general theorem giving a sufficient condition in order that almost all realizations of a process are continuously differentiable. Theorem 2.11 : Let X ()t ;β , tI∈ be a separable stochastic process. If for all tt, + h in the interval I the conditions of the Theorem 2.6 are satisfied and for all thtth−+,, in the interval I

c ()X ()()()th++−−;;2;βββ Xth Xt ≥ gh11() ≤ qh() (10)

where gh1 () and qh1 ( ) are even functions of h , non-increasing as h → 0 and such that

∞ ∞ nn− nn− ∑ 22g1 ()<∞ and ∑ 22q1 ()<∞ (11) n=1 n=1 then almost all realizations of a process X ()t ;β have continuous derivatives in I . Proof : See LOÈVE, M., (Probability Theory II). ■

114 CHAPTER 4 STOCHASTIC CALCULUS – PRINCIPLES AND RESULTS

4.2.4. Relation to Second-Order properties

For the sequel we assume that X ()t ;β , tI∈ is a second-order stochastic process. We shall now study the mutual connection between analytical properties of the sample functions of a process and its second-order properties.

Let us consider first integration of the process X ()t ;β , tI∈ . In general, may happens, the realizations of a m.s. integrable s.p. to be non-integrable. But, even if the realizations are integrable nevertheless one should differentiate the integrals,

I =()m.s.∫ X ()tdt ;β (12) I I * = ∫ X ()tdt;β (13) I where I is a m.s. of X ()t ;β , that is it is a random variable defined as a m.s. limit of appropriate approximating integral sums and I * is a function making a correspondence between the elementary event β ∈Ω and a Lebesque integral of the corresponding realization. In general, without assumption on measurability of a process, the integral I * needs not to be a random variable. The following theorem shows the relation between the two integrals

Theorem 2.12 : Let X ()t ;β , tI∈ be a measurable and m.s.-continuous stochastic process. Then the integrals I and I * exist and they are equal with probability one.

Proof : The existence of integrals I and I * follows directly from the assumptions. To show the equality with probability one it is sufficient to show that 2 Eβ II−⎯⎯⎯* n→∞→0 (14) ()n where In denotes an approximating sum of I . Of course, 22 EEEEββββI −=III***2 +⎡⎤ −2 II ()nn()[] ()⎣⎦⎢⎥ () n By virtue of eq. (1.25) Eβ I 2 → R t, t dt dt ()[]nXX∫∫ ()12 1 2 II× Making use of (4) one has 2 Eβ ⎡⎤I * = R t, t dt dt ()⎣⎦⎢⎥ ∫∫ XX ()12 1 2 II× 2 It can be easily shown that Eβ I *I approach the same quantity equal to Eβ ⎡⎤I * an ()n ()⎣⎦⎢⎥ n →∞. ■

Remark 2.13 : It can only be assumed that a process X ()t ;β , tI∈ is m.s.-continuous. Hence by virtue of Theorem 2.3 process X ()t ;β , tI∈ has stochastically equivalent measurable representation. ■

4.2. ANALYTICAL PROPERTIES OF SAMPLE FUNCTIONS 115

Let us consider now the questions associated with the continuity of a stochastic process X ()t ;β , tI∈ .

Theorem 2.14 : Let X ()t ;β , tI∈ be a separable second-order stochastic process, and there exists constants C > 0 and r > 0 such that, for every tt12, ∈ I 1+r RXX()tt22,,2,+− R XX () tt 11 R XX () tt 12 ≤− Ct 2 t 1 (15) then almost all realizations of X ()t ;β , tI∈ are continuous on I .

Proof : The proof is direct consequence of Kolmogorov Theorem with p = 2. ■

Remark 2.15 : Actually it is sufficient for the correlation function to have a second order ∂2 derivative RXX ()tt12, . The above conditions can be easily proved, using mean value ∂∂tt12 theorem, to be equivalent with condition (15). For the specific case of weakly stationary stochastic processes, there is the following

Theorem 2.16 : Let X ()t ;β , tI∈ be a weakly stationary process, m.s.-differentiable. Then its stochastically equivalent representation with continuous realizations exists.

Proof : Since X ()t ;β , tI∈ is a weakly stationary, m.s.-differentiable, stochastic process, we have

()RRXX()0 −≤ XX ()τ C τ Thus, the Kolmogorov condition a) takes the form ()p = 2 2 EXtβ ()()()+−τ ;;ββXt ≤ C τ So, we have proved the assertion. ■

We finish this section by presenting analogous results for differentiation of realizations of a stochastic process X ()t ;β , tI∈ .

Theorem 2.13 : Let X ()t ;β , tI∈ be a separable second-order stochastic process, and there ∂22n+ exists a finite derivative nn++11RXX ()tt12, . Then almost all realizations of X ()t ;β are n- ∂∂tt12 times differentiable.

Proof : See SOBCZYK, K., (Stochastic Differential Equations). ■

116 CHAPTER 4 STOCHASTIC CALCULUS – PRINCIPLES AND RESULTS

4.3. References

ASH, R.B., 2000, Probability and Measure Theory. Academic Press. ATHANASSOULIS, G.A., 2002, Stochastic Modeling and Forecasting of Ship Systems. Lecture Notes NTUA. CRAMER H. , LEADBETTER M. R.,1953, Stationary and Related Stochastic Process, Wiley NY. DOOB, J.L., 1953, Stochastic Processes. John Wiley &Sons. FRISTEDT, B. & GRAY. L., 1997, A Modern Approach to Probability Theory. Birkhäuser GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1996, Theory of Random Processes. Dover Publications. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1974, The Theory of Stochastic Processes I. Spinger. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1975, The Theory of Stochastic Processes II. Spinger. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1975, The Theory of Stochastic Processes III. Spinger. KLIMOV, G., 1986, Probability Theory and Mathematical Statistics. MIR Publishers. LOÈVE, M., 1977, Probability Theory II, Springer. PROHOROV, YU.V. & ROZANOV, YU.A., 1969, Probability Theory. Springer. PUGACHEV, V.S. & SINITSYN , 1987, Stochastic Differential Systems. John Wiley &Sons. SOBCZYK, K., 1991, Stochastic Differential Equations, Kluwer Academic Publishers. SOIZE, C., 1994, The Fokker-Planck Equation and its Explicit Steady State Solutions. World Scientific. SOONG, T.T., 1973, Random Differential Equations. Academic Press. VULIKH, B.Z., 1976, A Brief Course in the Theory of Functions of a Real Variable. MIR Publishers. YEH, J., 1995, Martingales and Stochastic Analysis. World Scientific.

Chapter 5

Probabilistic Analysis of the responses of Dynamical Systems under Stochastic Excitation

As we saw in Chapter 3 the characteristic functional gives a complete description of a stochastic process. In this chapter we shall see, how we can derive functional differential equations for the characteristic functional of a continuous time stochastic process whose sample functions satisfy a given Ordinary Differential Equation. This fact is of great importance, since we can formulate the full statistical problem into one equation. We must also note that, having this equation, all the statistical information of the problem can be recovered. Hence, form the functional differential equation we can, for example, derive the moments equations associated with the problem. The above formulation, although it contains the whole statistical information, it’s not yet of practical importance. The difficulty comes from the functional differential equation itself, since up to the present time very little has been done to solve this kind of equations. On the other hand, very few characteristic functions, can be generalized in infinite dimensional Hilbert spaces, in order to give representations of the characteristic functional. In Section 5.1 we introduce the notion of a stochastic differential equation and prove some general theorem concerning the existence and uniqness of solutions for equations of this kind. Both existence and uniqness is proved in the sense of probability measure. Moreover we study the necessary conditions for bounded of all moments of the probability measure that describes the system (ODE) response. Additionally we investigate the dependence of the solution of the system from a parameter inhered into the ODE. The analysis follows the steps of GIΚHMAN, I.I. & SKOROΚHOD, A.V., (The Theory of Stochastic Processes III). Section 5.2 deals with the formulation of functional differential equations for the joint characteristic functional that describes the joint probability structure of the excitation and the response. The formulation concerns stochastic systems described by nonlinear ODE’s with polynomial nonlinearities. We distinguish between the case of a m.s. continuous stochastic excitation and an orthogonal increment excitation. Before the general results we present two specific systems, the Duffing Oscillator and the Van der Poll Oscillator. In the next section 5.3 we prove how the general functional differential equation describing the probability structure of a system can be reduced to partial differential equations for the characteristic functions of various orders. Special cases of the above reduction are the general form of the Fokker-Planck equation as well as the Liouville equation. Of great importance are the partial differential equations proved at Section 5.3.4 that describes the probability structure of the system for the case of m.s. continuous excitation. In the last Section 5.5 we introduce the notion of the kernel characteristic functional that generalizes the corresponding idea of kernel density functions in infinite dimensional spaces. We prove that the Gaussian characteristic functional has the kernel property, based on existent

117 118 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

theorems for extreme values of stochastic processes. In Section 5.4.4 we present how the kernel characteristic functionals can be used efficiently for the solution of functional differential equations. The basic idea is the localization in the physical phase space domain (probability measure) expressed in terms of the characteristic functional. We must note that the Gaussian measures are of great importance for these constructions since every analytical calculation concerning infinite dimensional integrals holds for Gaussian measures. In this way we derive a set of partial differential equations that governs the functional parameters of the kernel characteristic functionals. Finally in Section 5.4.5 we make special assumptions to derive a more simplified set of equations. Moreover a simple nonlinear ODE is studied and numerical results are presented.

5.1. STOCHASTIC DIFFERENTIAL EQUATIONS – GENERAL THEORY 119

5.1. Stochastic Differential Equations – General Theory

In the present section we introduce the notion of a stochastic differential equation and prove some general theorems concerning the existence and uniqness of solutions of these equations. For this purpose it is necessary to generalize the notion of a stochastic integral introduced above. The following analysis is based on the approach of GIΚHMAN, I.I. & SKOROΚHOD, A.V., (The Theory of Stochastic Processes III).

5.1.1. General Problems of the Theory of Stochastic Differential Equations.

Assume that we are dealing with a motion of a system S in the phase space N and let N x()txtxtxt=()12 (),,, ()… N () denote the location of this system in at time t . Assume also that a displacement of system Σ located at point x during the time interval ()tt, +Δ tcan be represented in the form,

xxx()()(tt+−ΔΔδ tAttAt =,, +− )() x + (1)

Here, At()x, is in general a random function; the difference At()()xx,,+−Δ tAt characterizes the of an “external field of forces” at point x on Σ during the time period ()tt, +Δ t and δ is a quantity which is of higher order of smallness in a certain sense than the difference A()()x,,ttAt+−Δ x . If A()x,t as a function of t is absolutely continuous, then relation (1) can be replaced by the ordinary differential equation

dtx() = Att′()x(), (2) dt t

N Equation (2) defines the motion of Σ in for tt> 0 under the initial condition xx()t00= , while Att′()x, determines the “velocity field” in the phase space at time t. It is obvious that equation (2) cannot describe such as Brownian, i.e. motions which do not possess a finite velocity in the phase space or motions which possess discontinuities in the phase space. To obtain an equation which will describe a motion of systems of this kind it is expedient to replace relation (1) with an equation of an integral type. For this purpose we visualize that the time interval [tt0 , ] is subdivided into subintervals by the subdividing points tt12,,… , tn = t. It then follows from (1) that

nn−−11 x()tt−= x (01 )∑∑ Ax() ( ttii ),,+ − Ax()() tt ii +δ i ii==00 n−1 Since δi are of a small order, it is natural to assume that, ∑δi → 0 as n →∞. In this case the i=0 last equality formally becomes, t xxtt−= Ax ssds, (3) () (0 )∫ () () t0

120 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

and the expression t ∫ Ax()()ssds, t0 can be called a stochastic integral in the random field A()x,t along the random curve

x()sstt,,∈ [0 ]; the integral should be interpreted as the limit, in a certain sence, to be defined more precisely, of sums of the form

n−1 ∑ Ax()()ttii,,+1 − Ax()() tt ii i=0

Relation (3) is called a Stochastic Differential Equation and is written in the form

dtt xAx()==≥() () tt,, x() t00 x , tt 0

Under sufficiently general assumptions, for example, if A(x,t) is a martingale for each x ∈ N , one can assume that

A()xaxx,,,ttt=+ ()β () (4) where β()x,t as a function of t is a martingale and process ax( ,t) is representable as the difference of two monotonically nondecreasing natural processes. In this connection it makes sense to suppose that the right-hand side of equation (3) can be represented according to formula (4) and introduce further restrictions on functions ax(),t and β()x,t in various ways. For example, we may assume that the function ax(),t appearing in expression (4) is an absolute continuous function of t while β()x,t  as a function of t  is a local square integrable martingale. (Some more general assumptions concerning β()x,t are considered below). In what follows, equation (3) will be written in the form

tt xxtt=+ ax ssdssds,, +β x (5) () (0 )∫∫() () ()() tt00 or,

dtt xax()=+() () ttdt,,,β() x() tdt x() t00 = x

In the case when β()x()tdt,0= , equation (5) is called an ordinary differential equation (with a random right-hand side).

Often fields β()xxxx,,,,,,,tttt= ()ββ12 ()()… βN ()of the form

t M βγμxx,,,1,,tsdsnN==… (6) nnmm()∫ ∑ () () 0 m=1

are considered. Here μm ()sm,1,,= … M are local mutually orthogonal square integrable martingales and γnm (x, s) are random functions satisfying conditions which assure the 5.1. STOCHASTIC DIFFERENTIAL EQUATIONS – GENERAL THEORY 121

existence of corresponding integrals. In this case, the second integral in equation (5) may be defined as the vector-valued integral with components ttM βγμxxsds,,,1,,== ssd s n… N ∫∫nnmm()() ∑ ()() () m=1 tt00 and the theory of stochastic integrals described in Section 1 of Chapter 4 can be utilized. However, if we confine ourselves to functions β()x,t of type (6) a substantial amount of generality is lost. This can be seen from the fact that the joint characteristic of processes

βn ()x,t and βn ()y,t defined by formula (6) is of the form

t M β β F() tE≡⋅=()ββnn (xy,, t ) ( t ) γ nmnmmm( x , s ) γ ( y , sEdsds )() μμ () ⋅= () ββnn()()xy,,ii , ∫ ∑ 0 m=1 t M

= γγnm()()xy,,ssdFs nm () ∫ ∑ μμmm()ii, () 0 m=1 while in the general case it is given by a function Γ n ()xy,,t which for fixed t is an arbitrary nonnegative-defined kernel of arguments x and y,

Ι n ∑ Γ ()xyij,,tzz ij≥ 0 for all zi ∈ , nN=1, 2,… , , I =1, 2,… ij,1=

For example (for simplicity we consider here the one-dimensional case) let functions,

γnm()xx,tc=== nm () ,, t n 1,,,…… Nm 1,, M be nonrandom and μmm()twt= () be independent Wiener process. In this case the correlation function R()x,,y t of the field

t M β xx,,,1,,tcsdwsnN==… nnmm()∫ ∑ () () 0 m=1 equals t M R xy,,tE=⋅=β ββ x , t y , t c x , scsdsn y , , = 1,,… N nnnnmnm()() ()()∫ ∑ ()() 0 m=1

On the other hand, if we set βnn(xx,,tw)= ( t), where wtn ()x, is an arbitrary Gaussian field β with independent increments in t, its correlation function Rnnn()x,,y tEwtwt=⋅() ()()x ,y , is then an arbitrary nonnegative-definite kernel (for a fixed t). Thus, the restrictions, when considering stochastic integrals along a process x(t) imposed by fields of the type (6) lead to a substantial narrowing of the class of problems under consideration. Therefore it is expedient to introduce a direct definition and investigate properties of the stochastic integral t ∫ β()x()sds, t0 by interpreting it in the simplest case as the limit in probability of the sums n σβ()xx=−∑ () (sskk−−−111 ),, β() x() ss kk k=1 The sums σ are called integral sums.

122 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

It is appropriate to observe that remarks about the insufficient generality of random fields as given by relation (6) are not fully justified in the case when the stochastic processes are represented by equation (5), with ax( ,,tt)= α( x ) being a nonrandom function and β()x,t being a function with independent increments in t. Indeed, the increment Δx()t of a solution of equation (5) at each time t depends on x(t) and on the value of the field β()x,t at the point xx= ()t and is independent of the nature of the relationship between β()x,t and β()y,t at the point y ≠ x()t (provided the probabilistic characteristics of the field β()x,t as a function of x are sufficiently smooth). Therefore one may expect that solutions of equations (5) will be stochastically equivalent for any two fields ββ()x,,tt= 1 ()x and ββ()x,,tt= 2 ()x under the condition that the joint distributions of the sequence of vectors

N ()ββii()()xx,,tt12 ,,,… β in () x , t∀∈ x , ∀∈ n+

coincide for i =1 and i = 2 and that the fields, βi ()x,t possess independent increments in t. For example, let wx( ,t) be an arbitrary Gaussian field possessing independent increments in t,

β B()xxxx,,,,tEwtwt=⋅={}()kj () () {} Bt kj() and the functions Bkj ()x,t be differentiable d with respect to t, bt()xx,,= Bt(). Denote by σ()x,t a symmetric matrix such that kjdt kj 2 σ ()xx,,tbt= () and introduce independent Wiener process wtj (),1,, j= … N. Set t β xxw,,tsds= σ , w twtwtwt= ,,,… 1 ()∫ ( )() ()()12 () ()N () 0 Then tt Eβ ββxx,, t⊗= tσσ xx ,, s sds = b x , sds ()11() ()∫∫()() () 00 and one can expect that the solutions of the differential equations,

dt xax=+(),, t dt wx ( dt ) ,

dtdttdtttx =+ax(),,σ () x w ()

will be stochastically equivalent, although the fields wx(),t and β1 ()x,t in general are not.

Analogous observations can be made also in the case when β1 ()x,t is an arbitrary field with independent increments in t with finite moments of the second order. Assume that the increment ββ()()xx,,tt+−Δ t possesses the characteristic function Eittitβ ()exp()zx⋅+−⋅=ββ()() ,Δ zx , ⎛⎞tt++ΔΔ tt ⎜ 1 T ⎡⎤isxc⋅ () x,, u ⎟ exp⎜−⋅⋅+z b()xz , s ds ds e −−⋅ 1 i zcxu()() , s ,Π s , d u⎟ ⎜ 2 ∫∫∫⎣⎦⎢⎥⎟ ⎝⎠⎜ ttN (if β()x,t possess finite moments of the second order and is absolutely continuous in t one can reduce an arbitrary characteristic function of a process with independent increments to this 5.1. STOCHASTIC DIFFERENTIAL EQUATIONS – GENERAL THEORY 123

form). In this case, one can expect for sufficiently smooth functions ax(),t , bt()x, and cx(),,t u that solutions of the stochastic equations

dt xax=+(),, t dtβ ( x dt ) and dxax=++,, t dtσν x t dt cxu ,,, t dt d u t () ()∫ ( )( ) N will be stochastically equivalent. Here σ()x,t is a symmetric matrix, σ 2 ()xx,,tbt= (), t ⎡⎤ ν()tA, is a centered Poisson measure with Var⎣⎦νΠ()tA ,= ∫ () sAds , . 0 It is also clear that if the increments in t of the field β()x,t are dependent then the remarks above concerning the possibility of replacing the field β()x,t in equation (5) by a simpler field without restricting the class of obtained solution are no longer valid. The preceding outline of a definition of a stochastic differential equation is expedient to extend in yet another direction. At present, "feedback" systems play an important part in a number of scientific-engineering problems. For such systems the "exterior field of forces" acting on the system at a given time depends not only on the instantaneous location of the system in the phase space but also on its phase trajectory in "the past":

⎛⎞⎛⎞tt xxxtt+−ΔΔδ t =αα⎜⎜,, tt +−⎟⎟ x t + (7) ()()⎜⎜⎟⎟t ⎝⎠⎝⎠⎜⎜tt00 ⎛⎞t where α⎜x,,sstt⎟ ≥> is a family of random functionals with values in N , defined on a ⎜ ⎟ 0 ⎝⎠⎜ t0 N N certain space of functions x()sstt,,∈ [0 ] with values in . We now introduce the space DT of functions x(s) defined on (]−∞,t , continuous from the right with values in N possessing – at each point of the domain of definition – right-hand and left-hand limits and also the limit as s →∞. Let NN= and denote by θ ()tT≤ the mapping of N into N defined by the DDt0 t Dt D relation

()()(θt xxstss=+ ),0 ≤

Next let αα()xx,,,tt= (ω ) be a random function defined on D N ××Ω[]0,T . Relation (7) can be written as follows,

xxx()()()()tt+−ΔθΔθΔδ t =ααttt,, tt +− x tt ++ and equation (5) can be represented by the equation

tt xxt=+αβθθ x,,, s ds + x ds t > t (8) ()00∫∫ (ss ) ( ) tt00 Here it becomes necessary to define the process x(t) over the whole “past”, i.e., up to the time t0 . In this connection one should adjoin to equation (8) relation

xx()tttt=≤past () , 0 (9)

124 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

which will be called from now on the initial condition for the stochastic differential equation (8).

5.1.2. The Existence and Uniqueness Theorems of Stochastic Differential Equations

Let the probability space ()Ω, Uc()Ω , which is equipped with a current of σ −algebras N U t()Ω , tT∈[]0, . Let also the random functions a()φ,t and βφ(),t , with values in , N adopted to U t()Ω with φ ∈D , tT∈[]0, be given. For what follows we will use the concept of random time, so we have the following

Definition 1.1 : Let U t()Ω , tT∈[]0, be a current of σ −algebras. A function τω= f (),

ω ∈Ωτ ⊂Ω, with values in []0,T is called a random time on U t()Ω , tT∈[]0, or an

U t −random time, if {}τ ≤∈t U t . ■ Hence we can now give the definition of a solution of a stochastic differential equation.

Definition 1.2 : A random time τ , 0 <≤τ T , on U t()Ω , tT∈[]0, and a random process

x()t , tT∈[]0, , UUt ×−()[]0,T measurable and satisfying with probability 1 relations

xx()tttt=≤past () , 0 (10)

tt xxt=+αβθθ x,,, s ds + x ds t > t (11) ()00∫∫ (ss ) ( ) tt00 for each t ≤ τ , is called a solution of the stochastic differential equation

dxx=+αβ()θθtt,,, t dt ( x dt ) t > t0 (12) satisfying the initial condition

xx()tttt=≤past () , 0 ■ (13)

It is assumed that integrals in the r.h.s. of (11) are well defined, the first as a Lebesgue integral and the second as a stochastic integral. The random variable τ is called a lifetime of the process x(t) (i.e., of a solution of a stochastic differential equation). Equation (12) is called regular on []0,T provided it possess a unique solution on the whole time interval []0,T . (i.e., if a unique solution of equation (12) with τ = T exists).

We now introduce some general assumptions on functions a()φ,t and βφ(),t under which the r.h.s. of equality (11) will be well defined for a sufficiently wide class of processes x()t . Observe that there is no point to consider the most general classes of processes x(t) since the r.h.s. of equation (11) represents a process which possesses a continuous modification or a N modification with sample functions in DT and hence the process x(t) must also possess these properties. We shall first discuss the function a()φ,t . We introduce two sets of assumptions. 5.1. STOCHASTIC DIFFERENTIAL EQUATIONS – GENERAL THEORY 125

α1 a) The function aa()φφ,,,ss= (ω ) is defined on D N ××Ω[]0,T and is N UU()D ××Ω−()[]0,T Ut () measurable for sT∈[]0, . α1 b) The function aa()φφ,,,ss= (ω ) is defined on D N ××Ω[]0,T and is N UU()D ××Ω−()[]0,T Ut () measurable for sT∈[]0, . α1 c) For a fixed ω the family of functions {}a()i,,tt∈ [ 0, T ] of arguments φ is uniformly continuous on N relative to metric d . D D N α2 The function a()φ, s satisfies assumptions a1 provided D N , U ()D N and D []0,T are replaced by C N , U ()C N and C []0,T respectively. We say that the function a()φ,t is linearly bounded (relative to a uniform norm or a seminorm if in the succeeding inequalities the norm i can be replaced by the seminorm i ) provided a ∗ continuous monotonically nondecreasing process λ0 ()t exists adopted to U t ()Ω ,tT∈[]0, such that λ0 (t)<∞ with probability 1 and

bb a φλ,1tdt≤+φ tdt (14) ∫∫() ()0 () aa

We say that the function a()φ,t satisfies the Local Lipschitz condition (in a uniform norm or seminorm) if for any ν > 0 there exists a monotonically λν (t) exists adopted to

U t ()Ω ,tT∈[]0, such that

bb aaφψ,,1ttdt−≤+− φψλ tdt (15) ∫∫()() ()ν () aa for all φ and ψ satisfying φ ≤ N and ψ ≤ N . If we choose the process which is independent of ν in place of λν (t) then we shall refer to the process a()φ,t as one satisfying the uniform Lipschitz condition. The class of processes a()φ,t which satisfy condition a2, (14) and (15) will be denoted by c Sa ()λλ0 , ν . We denote by Sa ()λλ0 , ν the class of processes which satisfy a1 and the condition obtained from (14) and (15) when the uniform norm i is replaced by i in the ∗ corresponding inequalities. In the case when we shall be dealing with random functions a()φ,t satisfying only one of the inequalities (14) or (15), for instance (14), we shall write c a()φ,,t ∈Sa ()λ0 i and analogously in other cases. N N Note that φψtt= θ (ψ ∈D , tT∈[]0, ), with values in is a Borel function in argument t . n N Indeed if B is a cylinder in D with the basis B = ∏Bi in coordinates ()ss12,,… , snk , s≤ 0, i=1 then n {}tB::φφtii∈=∩{} ttsB() +∈ i=1

126 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

N Since the sets {}zzBZ:φ()∈=ii are Borel sets provided Bi ∈ are such, the set n {}tB:φtii∈=∩{} Zs −, where Z − s denotes the set {}zz: +∈ s Z, will also be a Borel set. i=1 Thus, if gt()φ, is a UU()D N ×−()[]0,T measurable function in the arguments ()φ,t where N N U ()D is the minimal σ −algebra generated by the cylinders in D , then gt()θtφ, will be a Borel function in argument t . Consequently if the field a()φ,t satisfies condition a1 or a2 then the integral

b a θ ψ,tdt ∫ ()t a exists with probability 1.

Next, if a()φ,,t ∈Sa ()λ0 i , then we have for 0 ≤<≤abT,

bb a θθλψψ,1tdt≤+ tdt (16) ∫∫()tt()∗ 0 () aa

The proof follows easily from the fact that θtψ ,0,tT∈[] is a continuous function with values in N with respect to a metric d in N , and a φ,t is a continuous function of argument D D N D () φ (uniformly in t ). c Analogously if a()φ,,t ∈Sa ()λ0 i , then

bb a θθλψψ,1tdt≤+ tdt (17) ∫∫()tt()0 () aa

If, however a()φ,,t ∈Sa ()i λν and ψψ12∨

bb b aa()θθθλψψ,,tdt−≤− () tdt() ψψ() tdt (18) ∫∫ttt12 ∫ 12∗ ν aa a

c and an analogous inequality is valid for the case a()φ,,t ∈Sa ()i λν . b As far as the integral β θ x,ds is concerned, conditions for its existence and its properties ∫ ()s a were discussed in the preceding chapter. We introduce the notation for the classes of fields βφ(),t analogous to the notation for the classes of functions a()φ,t introduced above. Namely we write βφ(),,t ∈Sβν()λλ0 (or c βφ(),,t ∈Sβν()λλ0 provided the field βφ( ,t) satisfies conditions β1 and β2), is linearly bounded, satisfies the local Lipschitz condition relative to a seminorm (norm) and the ′ dominating processes Λ0 ()t and Λν ()t are absolutely continuous with λΛ00(tt)= ( ) and ′ λ=Λνν()tt (). 5.1. STOCHASTIC DIFFERENTIAL EQUATIONS – GENERAL THEORY 127

Set t Aa()φφβφ,,tsdst=+∫ () () , t0 c and write A()φ,,t ∈S ()λλ0 ν ( A()φ,,t ∈S ()λλ0 ν ) provided

a()φ,,t ∈Sa ()λλ0 ν and βφ(),,t ∈Sβν()λλ0 c c a()φ,,t ∈Sa ()λλ0 ν and βφ(),,t ∈Sβν()λλ0

Let the process x(t) , tT∈[]0, , be a adopted to a current of σ −algebras U t ()Ω ,tT∈[]0, with sample functions belonging with probability 1 to D N []0,T . We complete the definition of N x()t for t ≤ 0 by setting x()tt= φ () for tt≤ 0 where φ(t) is a given function in D .

Assume that a()φ,t satisfies condition α1 a) and that βφ(),,t ∈Sβν()i λ . In what follows these condition will always be assumed to be valid. We now formulate the first basic theorems for the existence and uniqness of a solution in accordance with the operator A()φ,t . c Theorem 1.3 : Assume that A()φ,,t ∈S ()λλ0 ν (or that A()φ,,t ∈S ()λλ0 ν ). The stochastic differential equation (12) possesses in H ∗ (H c ) a unique solution which is defined for all tT∈[]0, . ■ Proof : For the proof see GIΚHMAN, I.I. & SKOROΚHOD, A.V., (The Theory of Stochastic Processes III. p.141) ■

c Theorem 1.4 : If A()φ,,t ∈S ()i λν (or A(φ,,t)∈S (i λν )) then a random time τ∞ exists on

U t ()Ω ,tT∈[]0, and a random process x()t adopted to U t ()Ω ,tT∈[]0, defined for t < τ∞ such that x(t) satisfies (12) for all t < τ∞ , and the sample functions of x(t) possess with probability 1 left – hand limits and are continuous from the right for all t < τ∞ . Moreover c ()τ∞ >=01. Proof : For the proof see GIΚHMAN, I.I. & SKOROΚHOD, A.V., (The Theory of Stochastic Processes III. p.142) ■

Theorem 1.5 : If under the condition of Theorem 1.3 one can set λ0 = C , then the solution of ∗ c equation (12) belongs to H 2 ( H 2 ) and ⎛⎞ β ⎜ 2 ⎟ 2 EtB⎜ supx() ⎟≤+() 1φ , (19) ⎝⎠⎜tT∈[]0, ⎟ ⎛⎞ β ⎜ 2 ⎟ 2 EstBh⎜ supxx()−≤+ ()⎟ () 1 φ (20) ⎝⎠⎜stth∈+[], ⎟ where B is a constant which depends on C , K and T only. For the proof see GIΚHMAN, I.I. & SKOROΚHOD, A.V., (The Theory of Stochastic Processes III. p.142) ■

128 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

5.1.3. Bounds on the Moments of Solutions of Stochastic Differential Equations

A very important issue for a solution of a stochastic differential equation is whether it possesses or not finite order moments. Based on theorems for the existence of moments, we can solve moment equations and have numerical results for the stochastic differential equation. The theorem presented below guarantees the existence of all moments for a very wide class of systems including Ito equations as well as equations with continuous excitation process.

Consider equation (12) - (13)

dxx=+αβ()θθtt,,, t dt ( x dt ) t > t0

xx()tttt=≤past () , 0 ■ satisfying the conditions of Theorem 1.3. Assume first that for a fixed φ , βφ( ,t) is a martingale process with continuous sample functions with probability 1. (GIΚHMAN, I.I. & SKOROΚHOD, A.V., The Theory of Stochastic Processes III. p.37, 144). Furthermore let the characteristic Ftk (), of the process β k ()φ,t be absolutely continuous with respect to the Lebesgue measure t Ftk = β φ, sds (21) ()∫ k ( ) 0 and

2 βλk ()φφ,1ss≤+()() (22)

For the existence of moments of the solution of an SDE we have the following

Theorem 1.6 : A process x(t) satisfying equation (12) - (13) and conditions (21) - (22) for tttN≤=τλN inf{} : () > possess moments of all orders. Proof : For the proof see GIΚHMAN, I.I. & SKOROΚHOD, A.V., (The Theory of Stochastic Processes III. p.144) ■ Corollary 1.7 : If

atCk ()φφ,1≤+() (23)

k 2 Et()β ()()φφ,|ΔΩ≤+ΔU t C() 1 t (24) where C is an absolute constant, then the solution of equation (12) - (13) possess finite moments of all orders for ttT∈[]0 , . ■

More extended results concerning the existence of all or of finite number moments and the corresponding conditions can be found at GIΚHMAN, I.I. & SKOROΚHOD, A.V., (The Theory of Stochastic Processes III. p.145). Moreover we must emphasize the issue of moment stabilities although we will not study it in the present work. For a detailed introductory discussion we refer to SOONG, T.T., (Random Differential Equations, p.256). Some interesting results for special cases of stochastic differential equations as well as stochastic partial differential equation can be found at the paper of KOTULSKI, Z. & SOBCZYK, K., (On the Moment Stability of Vibratory Systems with Random Parametric Impulsive Excitation). Another approach based 5.1. STOCHASTIC DIFFERENTIAL EQUATIONS – GENERAL THEORY 129

on the Fourier transform of the moment equations with respect to time have been studied. The transformed moments have been studied extensively in homogeneous turbulence theory, and we refer to BATCHELOR, G.K., (Theory of Homogeneous Turbulence) for a thorough discussion of the single-time equation.

5.1.4. Continuous dependence on a parameter of solutions of stochastic equations

Consider the equation of the form t ηηtt=+xAθ ,, sdstt ≥ (25) uu() ()∫ usu ( ) 0 t0

ηupast()tt=

()x past ()ttt, < 0 does not depend on u .

Theorem 1.8 : Assume that Axu (),,tCC∈S ( ) and, moreover, let 2 ≤ a) sup Et()xu () C ttT0 ≤≤ 2 b) lim supE xxu ()tt−=∀∈00 () 0 ttT[] , u→0 () ttT0 ≤≤ ⎛⎞tt+Δ 2 ⎜ ⎟ c) EtΔ−ΔΩ≤βψ(),,| βψ () tUU() E⎜ γψ() ,| sds () Ω⎟ ()utut0 ⎜⎟⎜ ∫ ⎟ ⎝⎠t and, for arbitrary N > 0 and ttT∈[]0 , , let ⎪⎪⎧⎫ limc ⎨⎬ supaauu()ψψγψ ,tt−+>=0 () ,() , tε 0 u→0 ⎩⎭⎪⎪ψ ≤N Then 2 lim supEtηηu ()−=0 () t 0 u→0 () ttT0 ≤≤

Proof : The variables ηu (t) possess finite moments of the second order since equations (25) -

(26) satisfy the conditions of Theorem 1.6. We represent the difference ηηu ()tt− 0 () in the form tt ηηtt−=+ σ tAAθθ η,, sds − η sds, uuusuus()00 () ()∫∫ ( ) ( ) tt00 where, tt σηηttt=−+xx Aθθ,, sdssds − A uu() ()0000 ()∫∫ us ( ) ( s ) tt00

130 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

It is easy to verify that ⎛⎞t 22⎜ 2⎟ Etηη()−≤ () t331 Et σ() ++ CTE2 ()⎜ θ ( ηη − ) ds⎟ = ()uu0 () ⎜∫ su0 ∗ ⎟ ⎜ ⎟ ⎝⎠t0

t t0 22 =+3Eσηη t C′ E su +++= su Kduds ()uu() ∫∫ ()()()0 () ts0 − 2 υηη=− Set uu()tEsssup () ()0 () . It follows from the last inequality that tst0 ≤≤ t 2 υσυtE≤+3 t sds uuu()() () ∫ () t0 Then it can be easily proved that 2 υσ≤ ′′ uu()tCsup E() () s tst0 ≤≤ where C′′ does not depend on u . Furthermore 22⎪⎧ σ ≤−+⎪ supEs()uu() 3⎨ sup Et()xx()0 () t ≤≤⎪ ≤≤ tst00⎩⎪ ttT 2 ⎛⎞⎛⎞T ⎜⎜ ⎟ ⎟ +−+Essds⎜⎜ aaθθηη,,⎟ ⎟ ⎜⎜∫ us()()000 s ⎟ ⎟ ⎜⎜ ⎟ ⎟ ⎝⎠⎜⎝⎠t0 ⎟ 2 ⎛⎞tt ⎜ ⎟ +−supEdsds⎜ βηθθ , βη , ⎟ ⎜ ∫∫us()()000 s ⎟ ttT0 ≤≤ ⎜ ⎟ ⎝⎠⎜ tt00⎟

=++3()I123II

By the condition, the quantity I1 → 0 . Next ⎛⎞t ⎜ 2 ⎟ I ≤−TE⎜ aaθθηη,, s s ds⎟ 2000⎜∫ us()() s ⎟ ⎜ ⎟ ⎝⎠t0 where the integrand is dominated by the quantity ⎛⎞t0 ⎜ 2 ⎟ 41CsuKdu2 ⎜ ++η ()()⎟ ⎜ ∫ 0 ⎟ ⎝⎠⎜ −∞ ⎟ which is independent of u and is integrable with respect to the measure ddsc × . On the other hand,

aaus()()θθηη000,,0ss−→ s in probability for each s and hence in measure ddsc × . Therefore I2 → 0 as u → 0. Finally 2 ⎛⎞tt⎛⎞ t ⎜ ⎟ ⎜ ⎟ I ≤−=4,EdsdsEsds⎜ βηθθ βη ,4,⎟ ⎜ γη θ⎟ 30000⎜ ∫∫us()() s⎟ ⎜ ∫ us()⎟ ⎜ ⎟ ⎜ ⎟ ⎜⎝⎠tt00⎟ ⎝⎠ t 0 and as in the case of quantity I2 it is easy to verify I3 → 0 as u → 0. ■

Remark 1.9 : We shall strengthen the assumption of Theorem 1.8 and assume that in addition to the conditions stipulated in the theorem 5.1. STOCHASTIC DIFFERENTIAL EQUATIONS – GENERAL THEORY 131

⎛⎞ ⎜ 2 ⎟ limEtt⎜ supxxu ()−=0 () ⎟ 0 u→0 ⎜ ⎟ ⎝⎠ttT0 ≤≤ In this case ⎛⎞ ⎜ 2 ⎟ limEtt⎜ supηηu ()−=0 () ⎟ 0 (27) u→0 ⎜ ⎟ ⎝⎠ttT0 ≤≤ The proof of this assertion is analogous to the proof of Theorem 1.8. ■

Theorem 1.10 : Consider the stochastic equation.

ddtttttuuxAxu==≤∈ u()()()[]θ s u,, x u x past ,00() 0, satisfying the conditions of Theorem 1.3 and for all N > 0 let ⎡⎤⎧⎫⎧⎫ ⎢⎥⎪⎪⎪⎪uu lim supcc⎨⎬⎨⎬ supλλ0 ()tp≥+ supN () tp ≥ = 0 p→∞ ⎢⎥⎪⎪⎪⎪ uttT⎣⎦⎩⎭⎩⎭⎪⎪⎪⎪00≤≤ ttT ≤≤ and also let condition c) of Theorem 1.8 be satisfied. Then, for any ε > 0 ⎪⎪⎧⎫ c ⎨⎬supxxu ()tt−>→0 () ε 0 as u → 0 . ⎪⎪≤≤ ⎩⎭⎪⎪ttT0 Proof : Let uu τλλpNu=≥≥≥∅=inf{}ttptptN :0 () , () ,x () () inf T p aauu()φφ,,tt= () for t < τ p p p au ()φ,0t = for t ≥ τ p , βφuup(),,tt=∧ βφ()τ t Aappφφβφ,,tsdst=+ p , uu()∫ () u () t0 Theorem 1.8 (or the corresponding remarks of this theorem) is applicable to equation

pp p ddtttttxAxuusu==≤()()()θ ,, x u x past , 0

p On the other hand, in view of Theorem 1.3 xxuu(tt)= ( ) with probability 1 for all t < τ p . Therefore for any ε > 0 ⎧⎫ ⎪⎪p cc⎨⎬sup xxuu()tt−>≤< () ετ{} p T ⎪⎪≤≤ ⎩⎭⎪⎪ttT0 Furthermore, ⎧⎫⎧⎫ ⎪⎪⎪⎪p ε cc⎨⎬⎨⎬supxxuuu()tt−>≤0 () ε sup xx() tt −>+ () ⎪⎪⎪⎪≤≤ ≤≤ ⎩⎭⎩⎭⎪⎪⎪⎪ttT00 ttT 3 ⎧⎫⎧⎫ ⎪⎪⎪⎪ppεε p +−>+−>≤cc⎨⎬⎨⎬supxx0 ()ttuu () sup xx() tt () ⎪⎪⎪⎪≤≤ ≤≤ ⎩⎭⎩⎭⎪⎪⎪⎪ttT0033 ttT ⎧⎫ ⎪⎪ppε ≤−>+

132 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

u u The uniform stochastic boundness of the processes λ0 (t) and λN ()t and Theorem 1.8 imply that one can first choose a sufficiently large value of p and N such that c {}τεp < 0 such that c ⎨⎬sup xx0 ()tt−>u () for u ∈[]0,δ ⎪⎪≤≤ ⎩⎭⎪⎪ttT0 3 ⎪⎪⎧⎫ c ⎨⎬sup xxu ()tt−><0 () εε. ■ ⎪⎪≤≤ ⎩⎭⎪⎪ttT0

5.2. HOPF-TYPE FDEs FOR THE CHARACTERISTIC FUNCTIONAL OF DYNAMICAL SYSTEMS 133

5.2. Hopf-type Functional Differential Equation (FDE) for the Characteristic Functional of Dynamical Systems

In this section we will derive the Functional Differential Equation for a wide class of autonomous stochastic dynamical systems with external forcing. To begin our discussion let us consider the special cases of a Duffing Oscillator and a van der Poll oscillator. Let ∞ ()Ω, Uc()Ω , be a probability space and CIc () be the separable, real, inner product (_ 2 inner product) space of infinitely differentiable functions with values in U ⊂ . ∞ ∞ U ()CIc () will denote the σ − field generated by the open subsets of CIc ().

5.2.1. Duffing oscillator

The Duffing oscillator has the following equation of motion

mx() t;;;βββ+++ cx () t kx () t ax3 () t ;; ββ = y () t (1)

⎡⎤⎡⎤ββ ββ ⎣⎦⎣⎦xt()()()()00;, xt ;= x 00 , x a.s. (2)

We assume that y()t ;β is a stochastic process defined onto the probability space

∞∞ ()Ω, Uc()Ω , with the associated probability space ()CIcc(),,Uc() CI () y , whose characteristic functional is Y y ()v . Using the results of the previous section we can conclude that there is a unique solution for the SDE (1)-(2). We desire to calculate the joint characteristic functional Y xy ()uv, . We first note that

u,exp,,, v=+ i x u i y v dx dy (3) Ycxy()∫∫ {} xy () ∞∞ CCcc ∞ where the integrals are over the infinite dimensional spaceCc . The brackets ii, denotes the inner product of the Hilbert space _ 2 , i.e. x, yxtytdt= ∫ () () . The Volterra derivative of I

δY xy ()uv, δY xy ()uv, Y xy ()uv, with respect to u(ξ) will be denoted as or simply . Direct δξu() δuξ calculations shows

δY xy ()uv, =+ixξ exp i x , u i y , v dx , dy (4) ∫∫ () {}c xy () δuξ ∞∞ CCcc

δY xy ()uv, Taking the nth ordinary derivative of with respect to ξ we will have δuξ

n n d δY xy ()uv, dx()ξ =+iixuiyvdxdyexp , , , (5) nn∫∫ {}c xy () duξδξ ∞∞ d ξ CCcc

Additionally we note that the third Volterra derivative of Y xy ()uv, with respect to uuu()ξξξ123,, () () is

134 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

δY xy ()uv, =+ixξξξ ix ixexp i x , u i y , v dx , dy (6) ∫∫ ()()()123{}c xy () δδδuuuξξξ ∞∞ 123 CCcc

Setting ξξξξ123=== we will have

3 δ Y xy ()uv, 3 =+⎡⎤ixξ exp i x , u i y , v dx , dy (7) 3 ∫∫⎣⎦() {}c xy () δuξ ∞∞ CCcc

Thus, we have

md2 δδδδYYYY()uv,,,, cd() uv k() uv a 3 () uv ⋅⋅xy +⋅⋅ xy +⋅ xy +⋅ xy = idξδ2 u id ξδ u i δ u i3 ⎡⎤3 ξξξδuξ ⎣⎦ (8) ⎧⎫2 ⎪⎪dx()ξξ dx() 3 =++++⎪⎪m c kxξξ axexp i x , u i y , v dx , dy ∫∫⎨⎬2 () () {}c xy () ∞∞⎪⎪ddξξ CCcc⎩⎭⎪⎪

Using the equation (1) governing the system we will have

2 3 mdδδδδYYYYxy()uv,,,, cd xy() uv k xy() uv a xy () uv ⋅⋅+⋅⋅+⋅+⋅=2 3 3 idξδ uξξξ id ξδ u i δ u i ⎡⎤δu ⎣⎦ξ (9) =+yixuiyvdxdyξ exp , , , ∫∫ () {}c xy () ∞∞ CCcc

Hence, the Functional Differential equation for the Duffing Oscillator will have the form

2 mdδδδYYYxy()uv,,, cd xy() uv k xy () uv ⋅⋅2 +⋅⋅ +⋅ + idξδ uξξξ id ξδ u i δ u 3 (10) a δδYYxy()uv,,1 xy () uv +⋅3 3 = iiv⎡⎤δu δ ξ ⎣ ξ ⎦

Concerning the initial conditions of the problem we note that

υδ()ii+= υδ′ (),0 Y xy ()12tt00 =++=expi x ,υδ()ii υδ′ () i y ,0() dx , dy (11) ∫∫ {}12tt00 c xy ∞∞ CCcc =+expixtυυ xt dx = υυ , ∫ ()10() 20 ()cYx () 012 ( ) ∞ Cc

5.2. HOPF-TYPE FDEs FOR THE CHARACTERISTIC FUNCTIONAL OF DYNAMICAL SYSTEMS 135

5.2.2. Van der Poll Oscillator

The van der Pol oscillator has the following equation of motion

mx() t;;1;;;ββ+−+= c() x () t2 x () t βββ kx () t y () t (12)

xt()()00;ββ= x and xt( 00;ββ) = x( ) a.s. (13)

Assuming that yt();β is a stochastic process defined onto the probability space

∞∞ ()Ω, Uc()Ω , with the associated probability space ()CCcc,,Uc()y , there is a unique joint Characteristic Functional

u,exp,,, v=+ i x u i y v dx dy Ycxy()∫∫ {} xy () ∞∞ CCcc

Similar calculations on the characteristic functional, as before, shows that

2 3 mdδδYYxy()uv,, cd xy() uv cd δδ YY xy() uv ,, xy () uv ⋅+⋅− +⋅=k idξ23 δξ u() i dtδδξut() u ( )2 id ξ δξ u() δξ u() t=ξ (14) ⎧⎫2 ⎪⎪dx()ξξ2 dx() =+−+⎪⎪m c xξξ1, kx eixu,,+ iyv dx dy ∫ ⎨⎬2 ()() () c xy () ∞∞⎪⎪ddξξ CCcc× ⎩⎭⎪⎪

By using the equation (12) we will have

md2 δδYY()uv,, c d3 () uv cd δ Y() uv , xy+⋅ xy − xy + idξδξ23 u() i dtδδξut() u ( )2 id ξδξ u() t=ξ (15)

δY ()uv, + +⋅kyedxdyxy = ξ ixu,, iyv , ∫ () c xy () δξu() ∞∞ CCcc×

Hence, the Functional Differential equation for the Duffing Oscillator will have the form

md2 δδYY()uv,, cd3 () uv cd δδ YY() uv ,,() uv ⋅+⋅−⋅+⋅=xy xy xyk xy idξ23 δξ u() i dtδδξut() u ( )2 id ξ δξ u() δξ u() t=ξ (16) 1 δY ()uv, = xy ivδ ξ

Concerning the initial conditions of the problem we note that

υδ()ii+= υδ ′ (),0 Y xy ()12tt00

=++=expi x ,υδ()ii υδ ′ () i y ,0c () dx , dy (17) ∫∫ {}12tt00 xy ∞∞ CCcc =+expixtυυ xt dx = υυ , ∫ ()10() 20 ()cYx () 012 ( ) ∞ Cc

136 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

5.2.3. Non-Autonomous Dynamical Systems under Stochastic Excitation I – Case of m.s. Continuous Excitation

∞,N Let ()ΩΩ,,Uc() be a probability space and CIc (), It=∞[)0 , be the separable, real, N ∞,N inner product space of infinitely differentiable functions with values in U ⊂ . U ()Cc ∞,N will denote the σ −field generated by the open subsets of CIc (). Let the dynamical system

x=Fxy ()tttt;,;,;βββ() ()() a.s. (18) ( Σ 1 ) x=x()t00;ββ (), a.s. (19)

where, a) y()ii;:I ×Ω → M is a mean square continuous stochastic process defined onto the probability space ()ΩΩ,,Uc() with the associated probability space

∞∞,,MM ()CCcc,,Uc()y , and with characteristic functional Y y .

N b) x0 ()i:Ω→ is a random variable with the associated probability space MM () ,,Uc() x with characteristic function Y 0 . 0

c) F ()iii,, :I ×× NM → N is a measurable, deterministic function that can be expanded into a series of the form

KL ⎪⎪⎧⎫(kl, ) F ()tt,,x y =∇∇klFx ( , ,yyy )⎨⎬⎪⎪ ⊗−−()()xx (20) ∑∑{}yx 00⎪⎪ 0 0 kl==00 ⎩⎭⎪⎪

(kl, ) where ∇∇mlis the tensor product ∇ ⊗…… ⊗∇ ⊗∇ ⊗∇ , ⊗−()()yy xx − is the y x yyxx 00 m l tensor product ()yy−⊗⊗−⊗−⊗⊗−…… ()()() yy xx xx and the symbol denotes 0000 kl

the inner product between the two tensors. For the sequel we assume, for simplicity that x0 = 0

and y0 = 0 .

Theorem 2.1 : The problem Σ1 is equivalent to the functional differential equation

11d KL ⎡⎤∇=∇∇∇∇()uv,,,,kl F (t x y ) l k ()uv (21) ⎣⎦⎢⎥ut()YYxy∑∑{} y x00 kl+ ut() vt () xy idtkl==00 i

υυ⋅=δ ,0 (), υ ∈ N (22) Y xy ()t0 Y 0

∞,N ∞,M for every u ∈ CIc (), v ∈ CIc () and tI∈ , It=∞[)0 , . Where Y xy is the joint characteristic functional of the s.p.’s x()t ;β , y()t ;β , tI∈

⎛⎞⎛⎞ δδYYxy xy δ Y xy ⎟ δδYYxy xy δ Y xy ⎟ ∇= ⎜ ,,,… ⎟ and ∇= ⎜ ,,,… ⎟ u()ξ ⎜ ⎟ v()ξ ⎜ ⎟ ⎝⎠⎜δξδξuu12() () δ uN () ξ⎟ ⎝⎠⎜δξδξvv12() () δ vM () ξ⎟ 5.2. HOPF-TYPE FDEs FOR THE CHARACTERISTIC FUNCTIONAL OF DYNAMICAL SYSTEMS 137

are the Volterra gradients of the functional Y xy with respect to every argument uv, and lk ∇∇ut() vt () is the operator defined by the (lk, )-tensor product of the operators ∇u()ξ and ∇v()ξ , i.e. ∇∇lk =∇⊗ …… ⊗∇⊗∇ ⊗∇ With υ⋅=δυυυδ()i…i (,,, ) ⋅− (t ) we ut() vt () ut () ut () vt () vt () tN0 12 0 lk denote the vector valued Dirac delta function centred at zero.

Proof : Let c be the joint probability measure. Multiplying every equation

xtnn();,;,;βββ=x Ft() ()() t y t nN=1, 2,… ,

∞∞,,NM with exp{}iixu ,+ y ,v and integrating over CCcc× with respect to the probability measure c we will have for every nN= 1,2,… , ,

x t;exp,β ixu+ i yv , dx , dy = ∫ n () {}c xy () CC∞∞,,NM× cc (23) F t,,xy exp i xu ,+ i yv , dx , dy ∫ n (){}c xy () ∞∞,,NM CCcc×

For the l.h.s of the above equation we can easily notice that,

1 d δY ()uv, x t;exp,β ixu+ i yv , dx , dy = xy (24) ∫ n () {}c xy () ∞∞,,NM idtδ un () t CCcc×

For the r.h.s. we have using (20)

F t,,xy exp i xu ,+= i yv , dx , dy ∫ n (){}c xy () ∞∞,,NM CCcc×

KL ⎪⎪⎧⎫()kl, =∇∇⊗+klF() t,0,0⎨⎬⎪⎪yx exp{} i x , u i y , v () dx , dy = ∫ ∑∑{}yxn ⎪⎪ c xy ∞∞,,NMkl==00 CCcc× ⎩⎭⎪⎪

KL ∂kl+ Ft(),0,0 =+=n xyijexpi xu , i yv , dx , dy ∫ ∑∑ ∑ {}c xy () ∞∞,,NMkl==00()ij, ∈×IINM ∂∂xyij CCcc× ij=∧lk =

KL1 ∂kl+ Ft(),0,0 =+=n iiidxdykl+ xyijexp xu , yv , , ∑∑kl+ ∑ ∫ {}c xy () kl==00i ()ij, ∈×IINM ∂∂xyij ∞∞,,NM CCcc× ij=∧lk =

KL kl+ kl+ 1 ∂ Ft(),0,0 δ Y xy ()uv, =⋅=n ∑∑kl+ ∑ kl==00itt()ij, ∈×IINM ∂∂xyijδδ u i() v j () ij=∧lk =

KL kl1 l k =∇∇∑∑{}yxFtn (),0,0kl+ ∇∇ut() vt ()Y xy()uv , . kl==00 i

Where we have used the notation,

ii12 iN ∂=∂⋅∂xi x12xx… ⋅∂N

i ii12 iN x =⋅⋅⋅x12xx… N

138 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

Hence, we have proved the equality, 11d δY ()uv, KL xy =∇∇kl ∇∇ l k ∑∑{}yxFtn (),0,0kl+ ut() vt ()Y xy()uv , (25) idtδ un () tkl==00 i

or equivalently,

11d KL ⎡⎤∇=∇∇∇∇()uv,,,,kl F (t x y ) l k ()uv (26) ⎣⎦⎢⎥ut()YYxy∑∑{} y x00 kl+ ut() vt () xy idtkl==00 i

Concerning the initial conditions, we have by direct calculation of the l.h.s.of equation (22)

υυ⋅=δδ()i ,0 expidxdyx , ⋅() , = Ycxy()tt00∫ {} xy ∞∞,,NM CCcc×

⎪⎪⎧⎫N ==exp⎨⎬⎪⎪i xυ () dx , dy ()υ ■ (27) ∫ ⎪⎪∑ nnc xy Y 0 ∞∞,,NM ⎪⎪n=1 CCcc× ⎩⎭ Example 2.2 [One – Dimensional Non-Linear Oscillator] : To illustrate the above Theorem we will study the special case of the one-dimensional non-linear oscillator, i.e. the system

mxt();;,;;ββββ+= f() xt ()() xt yt() a.s. (28) ( Σ1a ) ⎡⎤⎡⎤ββ ββ ⎣⎦⎣⎦xt()()()()00;, xt ;= x 00 , x a.s. (29) where, we have made the same hypothesis as in Theorem 5.1. We also assume that f ()ii, can be expanded in the form NM ij f ()xx12, = ∑∑ axxij 12 (30) ij==00 Transforming the above equation into a state-space form we have

x12()txt;;ββ= ()

mx221() t;;,;;ββββ+= f() x ()() t x t y() t ⎡⎤⎡⎤ββββ ⎣⎦⎣⎦xt10(;, ) xt 20 ( ; )= x 0 () , x 0 ()

By direct application of the Theorem 2.1 we will have the two functional differential equations

d δδYY()uuv,,() uuv ,, xy12= xy 12 dtδδ u12() t u() t

mdδδYY()uuv,,NMa nm+ () uuv ,, xy12+⋅nm xy 12 = ∑∑ nm+ idtδδξδξδξδξ u2212111() tnm==00 i u()…… u ()nn u()++ u ( nm ) ξ j =t

1 δY xy ()uuv12,, = ivtδ () Substituting the first Functional Equation into the second we have, 5.2. HOPF-TYPE FDEs FOR THE CHARACTERISTIC FUNCTIONAL OF DYNAMICAL SYSTEMS 139

⎛⎞ 2 NM ⎜ n nm+ ⎟ mdδδYYxy()uv,,a ⎜ d xy () uv ⎟ ⋅+⋅⋅nm ⎜ ⎟ = 2 ∑∑ nm+ ⎜ ⎟ idξδξ u() == i d ξ…… d ξδξδξ u() u ξ =t ⎟ ξ=t nm00 ⎜ 11nnm()+ j ⎟ ⎝⎠⎜ jnm=+1,… ⎟

1 δY xy ()uv, =⋅ ivtδ () ∞ ∞ for every uC∈ c () I,v ∈ CIc () and tI∈ with υδ()ii+= υδ′ (),0() υ , υ , ()υυ, ∈ 2 YYxy ()12tt00 01212

140 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

5.2.4. Non-Autonomous Dynamical Systems under Stochastic Excitation II – Case of Independent Increment Excitation

∞,N Let ()ΩΩ,,Uc() be a probability space and CIc () be the separable, real, inner product N ∞,N space of square integrable functions with values in U ⊂ . U ()CIc () will denote the ∞,N σ − field generated by the open subsets of CIc ( ). Let the dynamical system

dtttx=Fx();,;,;;ββ() tt () dttt+ Gxy()() ββ dt() a.s. (31) ( Σ 2 ) x=x()t00;ββ (), a.s. (32)

where, a) y()ii;:I ×Ω → M is a stochastic process with independent increments defined onto the probability space (ΩΩ,,Uc( ) ) with the associated probability space

∞∞,,MM ()CCcc,,Uc()y , and with characteristic functional Y y .

N b) x0 ()i:Ω→ is a random variable with the associated probability space NN () ,,Uc() x with characteristic function Y 0 . 0

c) F ()ii,:T ×→ NN is a measurable, deterministic function that can be expanded into a series of the form L l F ()tt,,xFxxx=∇l ( ) ⊗−() (33) ∑{}x 00{} l=0

d) G()ii,:T ×→ NNM × is a measurable, deterministic function that can be expanded into a series of the form

K k Gx()tt,,=∇k Gx ( ) ⊗−() xx (34) ∑{}x 00{} k=0

l where ∇l is the tensor product ∇⊗∇… , ⊗−x x is the tensor product x xx ()0 l ()x −⊗⊗−xxx… () and the symbol denotes the inner product between the two tensors. 00 l

For the sequel we assume, for simplicity that x0 = 0 and y0 = 0 .

Remark 2.3 : We must note that since y()ii;:I ×Ω → m is a stochastic process with independent increments, it has nowhere differentiable sample functions. This explains the fact that the equation describing the dynamical system is in terms of differentials. ■

5.2. HOPF-TYPE FDEs FOR THE CHARACTERISTIC FUNCTIONAL OF DYNAMICAL SYSTEMS 141

Theorem 2.4 : The problem Σ2 defined above, is equivalent with the functional differential equation

L 11⎡⎤ll dtdtt ∇=∇∇+YYxy()uv,,0,{} x F () xy () uv ii⎣⎦⎢⎥ut() ∑ l ut() l=0 (35) K 1 +∇kkGuv()td,0 ∇⎡ ∇ () , ⎤ ∑{}xxyk+1 ut()⎣⎢ t vt ()Y ⎦⎥ k=0 i

υυ⋅=δ ,0 (), υ ∈ N (36) Y xy ()t0 Y 0

∞,N ∞,M for every u ∈ CIc (), v ∈ CIc () and tI∈ . Where Y xy is the joint characteristic functional of the s.p.’s x()t ;β , y()t ;β , tI∈

⎛⎞⎛⎞ δδYYxy xy δ Y xy ⎟ δδYYxy xy δ Y xy ⎟ ∇= ⎜ ,,,… ⎟ and ∇= ⎜ ,,,… ⎟ u()ξ ⎜ ⎟ v()ξ ⎜ ⎟ ⎝⎠⎜δξδξuu12() () δ uN () ξ⎟ ⎝⎠⎜δξδξvv12() () δ vM () ξ⎟

l are the Volterra gradients of the functional Y xy with respect to every argument uv , and ∇ut() is the operator defined by the l -tensor product of the operator ∇u()ξ , i.e. ∇=∇⊗⊗∇l … With υ⋅=δυυυδ−()(),,,…i ⋅ t we denote the vector valued ut() ut () ut () tN0 12 0 l Dirac delta function centred at zero. ■

Proof : Let c xy be the joint probability measure. Multiplying every equation

M dxtn() t;,;ββ=x F n() t () t dt+ ∑ G nm() t ,;; x() t ββ dy tm() t nN= 1, 2,… , m=1

∞∞,,NM with exp{}iixu ,+ y ,v and integrating over CCcc× with respect to the probability measure c xy we will have for every nN=1, 2,… , ,

dx texp ixu ,+ i yv , dxdy , = ∫ tn() {}c xy () ∞∞,,NM CCcc× =++F t,exp,xxuyv dt i i , dx , dy ∫ n () {}c xy () ∞∞,,NM CCcc× M ++Gt,;exp,,xxuyv dytβ i i dxdy , ∫ ∑ nm() t m () {}c xy () ∞∞,,NMm=1 CCcc×

For the l.h.s of the above equation we can easily notice that,

1 δY xy ()uv, dx t;exp,β ixu+ i yv , dxdy , = d ∫ tn() {}c xy () t ∞∞,,NM iutδ n () CCcc×

For the first term of the r.h.s. we have using (33)

F t,exp,xxuyv i+= i , dx , dy ∫ n () {}c xy () ∞∞,,NM CCcc×

142 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

L l =∇⊗l F() t,0 xxuyv exp i , + i ,() dx , dy = ∫ ∑{}xxyn {} {}c ∞∞,,NMl=0 CCcc×

L ∂l Ft(),0 =+=n xxuyvi expi , i , dx , dy ∫ ∑∑ {}c xy () N ∂x CC∞∞,,NM× l=0 i∈I i cc i =l

L 1 ∂l Ft(),0 =+=n iiil xxuyvi exp , , dxdy , ∑∑l ∫ {}c xy () i N ∂x l=0 i∈I i CC∞∞,,NM× i =l cc

LLll 11∂ Ftn (),0δ Y ()uv , =⋅=∇∇ll ∑∑l ∑{}xxyFtn (),0l ut()Y ()uv , . ll==00itii∈I N ∂xuiiδ () i =l

Where we have used the notation,

ii12 iN ∂=∂⋅∂xi x12xx… ⋅∂N

i ii12 iN x =⋅⋅⋅x12xx… N

For the second term of the r.h.s. we have using (34)

M G t,exp,,,xxuyv t d y t i+= i dx dy ∫ ∑ nm()() t m () {}c xy () ∞∞,,NMm=1 CCcc×

MK k =∇⊗k G() t,0 xxuyv() t d y() t exp i , + i ,() dx , dy = ∫ ∑∑{}x nm{} t m {}c xy ∞∞,,NMmk==10 CCcc×

MK k ∂ Gt(),0 i =+=nm xxuyvt d y texp i , i , dx , dy ∫ ∑∑∑ ()tm () {}c xy () N ∂x CC∞∞,,NM× mk==10i∈I i cc i =k

MK k 1 ∂ Gt(),0 i =+=nm itdytik+1 xxuyvexp , i , dxdy , ∑∑k+1 ∑ ∫ ()tm () {}c xy () i N ∂x mk==10 i∈I i CC∞∞,,NM× i =k cc MK 1 ∂kGt(),0 δ k ⎡ δY ()uv, ⎤ ==nm ⎢d xy ⎥ ∑∑k+1 ∑ ⎢ t ⎥ N ∂ δδ mk==10itvti∈I xuii()⎣⎢ m () ⎦⎥ i =k

MK ⎡ δ uv, ⎤ kk1 ⎢ Y xy ()⎥ =∇{}x Gtnm(),0 ∇ d t . ∑∑ k+1 ut()⎢ δ ⎥ mk==10 ivt⎣⎢ m () ⎦⎥

Hence we have proved the equality,

11δY ()uv, L xy =∇ll ∇ + ddtFttn∑{}xxy(),0l ut()Y ()uv , iutδ n () l=0 i

MK ⎡ δ uv, ⎤ kk1 ⎢ Y xy ()⎥ +∇{}x Gtnm(),0 ∇ d t ∑∑ ivtk+1 ut()⎢ δ ⎥ mk==10 ⎣⎢ m () ⎦⎥ or equivalently, 5.2. HOPF-TYPE FDEs FOR THE CHARACTERISTIC FUNCTIONAL OF DYNAMICAL SYSTEMS 143

L 11⎡⎤ll dtdtt ∇=∇∇YYxy()uv,,0,{} x F () xy () uv + ii⎣⎦⎢⎥ut() ∑ l ut() l=0 K 1 +∇kkGuv()td,0 ∇⎡ ∇ () , ⎤ ∑{}xxyk+1 ut()⎣⎢ t vt ()Y ⎦⎥ k=0 i

Concerning the initial conditions, we have by direct calculation of the l.h.s. of equation (36)

υυ⋅=δδ,0 expidxdyx , ⋅() , = Ycxy()tt00∫ {} xy ∞∞,,NM CCcc×

⎪⎪⎧⎫N ==exp⎨⎬⎪⎪i xυ () dx , dy ()υ ■ (37) ∫ ⎪⎪∑ nnc xy Y 0 ∞∞,,NM ⎪⎪n=1 CCcc× ⎩⎭

Example 2.5 [Linear Filtering System] : Let us consider as a special case of the above theorem the a linear time invariant system, i.e. a system of the form,

dtttx=FxGw();;ββ⋅+⋅ () t dtdt () ; β a.s. (38)

x=x()t00;ββ (), a.s. (39) where w()ii;:I ×Ω → M is the standard Wiener process and F ∈ NN× and G ∈ NM× The above system is very essential for linear system theory and linear filter design, since we have explicit expression for the spectrum of the response in terms of the excitation spectrum (Wiener-Khinchin relations). For a detailed discussion we refer to PUGACHEV, V.S. & SINITSYN, (Stochastic Differential Systems). Now, let us derivate the associated functional differential equation for the joint characteristic functional of the response and the excitation. By direct application of Theorem 2.4 we have the functional differential equation

ddtd⎡⎤∇=⋅∇+⋅∇()uv,, Fl () uv G () uv , (40) tt⎣⎦⎢⎥ut()YYxy ut() xy vt() Y xy

N Y xy ()υυ⋅=δ−()i t00,0 Y (), υ ∈ (41)

∞,N ∞,M for every u ∈ CIc (), v ∈ CIc () and tI∈ . Where Y xy is the joint characteristic functional of the s.p.’s x()t ;β , w()t ;β , tI∈ . ■

144 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

5.3. Reduction of the Hopf-Type Functional Differential Equation for Special Cases

In this section we will study, how the general Functional Differential Equation can be reduced to partial differential equations when the external excitation fulfils special conditions. We will start our discussion with the case of deterministic systems (absence of stochastic excitation), with random initial conditions. Using the Hopf-type FDE we will prove Liouville Equation for the characteristic function of the response.

5.3.1. Systems with Random Initial Conditions – Liouville Equation

∞,N Let ()ΩΩ,,Uc() be a probability space and CIc () be the separable, real, inner product (_ 2 inner product) space of infinitely differentiable functions with values in U ⊂ N . ∞,N ∞,N U ()Cc will denote the σ −field generated by the open subsets ofCIc (). Let the dynamical system

x=Fx ()ttt;,;ββ() () a.s. (1) ( Σ3 ) x=x()t00;ββ (), a.s. (2)

where, N a) x0 ()i:Ω→ is a random variable with the associated probability space NN N () ,,Uc() x with characteristic function Y 0 ()υυ, ∈ . 0 b) F ()ii,:I ×→ NN is a measurable, deterministic function that can be expanded into a series of the form

L l F ()tt,,xFxxx=∇l ( ) ⊗−() (3) ∑{}x 00{} l=0

For the sequel we assume, for simplicity that x0 = 0 . For the case where an explicit solution of the above system of equations can be found, the solution would have the following form

ββ⎡⎤ x=gx()tt;,⎣⎦0 () (4)

The mapping (4) characterize at arbitrary time t the state of a system which has an initial state described by the random variable (2). In other words, these functions represent at time t a transformation of the random vector (2). It is clear that the probability distribution of the response can be obtained by use of the theorem of transformations of random vectors.

Let us assume that the transformation defined by (4) is continuous with respect to x0 ()β (for

each t), has continuous partial derivatives with respect to x0 ()β , and defines a one-to-one mapping. Then, if the inverse transformation is written as

ββ⎡⎤ x=hx0 ()⎣⎦tt,; ( ) (5)

the probability density of the response takes the form

5.3. REDUCTION OF THE HOPF-TYPE FUNCTIONAL DIFFERENTIAL EQUATIONS FOR SPECIAL CASES 145

∂hx[]t, ftftt()xhx,,,=⋅() [] (6) 0 ∂x

For the case where no explicit solution of the (1) can be found, we return to the Hopf-type FDE associated with the stochastic initial value problem described above. Using Theorem 2.1 we have

11d L ⎡⎤∇=∇∇()uFll (t,0 ) () u (7) ⎣⎦⎢⎥ut()YYxx∑{}l ut() x idtl=0 i

N Y xy ()()υυυ⋅=δ00,0Y , ∈ (8)

∞,N for every u ∈ CIc () and tI∈ . Where Y x is the characteristic functional of the s.p. x()t ;β , tI∈ . For the case described above the following theorem holds

Theorem 3.1 [Liouville equation] : Let the dynamical system Σ3 described by (1)-(2). Then the FDE (7)-(8) which describes the system can be reduced to the equation

∂φ ()t,υ NLυ x +∇∇=∈∈n llφ υυ N ∑∑ l+1 {}xxFtn (),0{}υ () t , 0 , , t I (9) ∂tinl==10

N φx ()0,υυυ=∈Y 0 () , (10) where φ()t,υ is the characteristic function of the response. ■

Proof : Let’s consider the component wise form of the FDE (9), used for the derivation at

Theorem 2.1.

11d δδYY()uuL ∂l Ft(),0 l () xxn = … =⋅∑∑l , nN1, 2, , idtδδ un () tl=0 ii∈I N ∂xuii() t i =l

Setting to the system of functional differential equations u =⋅−()()υυ12,,…i υN δ t and substituting uxu= expidx , we have Ycxx() ∫ {}() ∞,N Cc

⎪⎪⎧⎫N xt ()exp⎨⎬⎪⎪ i xt ()υ ( dx )= ∫ nkk⎪⎪∑ c x ∞,N ⎪⎪k=1 Cc ⎩⎭ Lnl 1 ∂ Ft(),0 i ⎪⎪⎧⎫ =⋅n itl ⎡⎤xxexp⎪⎪ ixtυ d ⇔ ∑∑l ∫ ⎣⎦()⎨⎬ ∑kk ()c x ( ) i N ∂x ⎪⎪ lk==01i∈I i C∞,N ⎩⎭⎪⎪ i =l c

Where we have used the notation,

ii12 iN ∂=∂⋅∂xi x12xx… ⋅∂N

i ii12 iN x =⋅⋅⋅x12xx… N

Multiplying both sides with iυn and summing up for all nN=1,2,… , , we will have,

NN⎪⎪⎧⎫ ixtυυ ()exp⎨⎬⎪⎪ i xt () ( dx )= ∑∑∫ nn⎪⎪ k kc x nk==11∞,N ⎪⎪ Cc ⎩⎭ 146 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

NLl n 1 ∂ Ft(),0 i ⎪⎪⎧⎫ =⋅iitixtdυυn l ⎡⎤xxexp⎪⎪ ⇔ ∑∑∑nkkl ∫ ⎣⎦()⎨⎬ ∑ ()c x ( ) i N ∂x ⎪⎪ nl==10i∈I i C∞,N ⎩⎭⎪⎪ k = 1 i =l c

But NN⎪⎪⎧⎫∂φ ()t,υ ixtυυ ()exp⎨⎬⎪⎪ i xt () ( dx )= x ∑∑∫ nn⎪⎪ k kc x nk==11∞,N ⎪⎪ ∂t Cc ⎩⎭ and NLl n 1 ∂ Ft(),0 i ⎪⎪⎧⎫ iitixtdυυn ⋅=l ⎡⎤xxexp⎪⎪ ∑∑∑nkkl ∫ ⎣⎦()⎨⎬ ∑ ()c x ( ) i N ∂x ⎪⎪ nl==10i∈I i C∞,N ⎩⎭⎪⎪ k = 1 i =l c

NL ll iυn ∂∂Ftn (),0φx () t ,υ =⋅∑∑l ∑ nl==10i i∈I N ∂∂xiiυ i =l

Hence, we have proved the equality,

NL ll ∂∂∂φφxx()tFtt,,0,υυiυn n () () =⋅∑∑l ∑ (11) ∂∂∂tinl==10 i∈I N xiiυ i =l or equivalently,

∂φ ()t,υ NLυ x +∇∇=n ll φ υ ∑∑ l+1 {}xxFtn (),0{}υ () t , 0 ■ (12) ∂tinl==10

Remark 3.2 : Taking the Fourier transform of the equation (11) or (12) we can calculate that the equation governing the probability density of the response is

∂ft(), x N ∂ x +=∈∈⎡⎤ft,,0,,xx Ft t I x N (13) ∑ ⎣⎦x ()()n ∂∂txn=1 n which is the Liouville equation from the general theory of dynamical systems. It is worth emphasizing the fact that equations of the form (13) expresses, in general conservation of some quantities in time. For example in continuum mechanics, such equation is known as the continuity equation, the equation of conservation of mass or the transport equation. Here, the Liouville equation expresses the conservation of probability during a motion of the dynamical system. ■

5.3. REDUCTION OF THE HOPF-TYPE FUNCTIONAL DIFFERENTIAL EQUATIONS FOR SPECIAL CASES 147

5.3.2. Moment Equations from the FDE

As we mention at the beginning a very important property of the formulation based on the characteristic functional, is that we can recover all the statistical information of the problem. In this section we will show, how we can derive the set of moment equations, associated with the stochastic differential equation, from the functional differential equation. Moreover, based on this analysis we will derive an equivalent set of integrodifferential equations with the functional differential equation derived at Theorem 2.1. As we prove at the previous section the stochastic differential equation Σ1 is equivalent to the FDE

11d KL ⎡⎤∇=∇∇∇∇()uv,,0,0,kl F (t ) l k () uv (14) ⎣⎦⎢⎥ut()YYxy∑∑{} y xkl+ ut() vt () xy idtkl==00 i N Y xy ()()υυ⋅=δ00,0 Y , υ ∈ (15)

∞,N ∞,M for every u ∈ CIc (), v ∈ CIc () and tI∈ . First we recall some important results concerning the polynomial expansion of functionals.

∞∞,,NM Theorem 3.3 [Taylor’s formula with integral reminder]: Let Z :h ={}CCcc×→ be a functional of class V r+1 . If the interval []aa, + h is contained in h then

1 ZZZZ(ah+= ) () a +δδ ()[] ah +2 ()[ ahh, ] ++… 2 1 r 1 ()1−t +++δδrrZZ()[ah,,…… h ] +1 ()[] atdt hh ,, h NN!!∫ r times0 r+1 times

Proof : For the proof we refer to CARTAN, H., (Differential Calculus. p.70). ■

Theorem 3.4 [Taylor’s formula with Lagrangian reminder]: Let the V r+1 functional ∞∞,,NM Z :h ={}CCcc×→ . If

δ r+1Z ()aa≤∈M for h then r+1 1 h ZZZ(ah+− ) () a −δδ ()[] ah −−……r Z()[ ah,, h ] ≤M Nr!1! ()+ r times

Proof : For the proof we refer to CARTAN, H., (Differential Calculus. p.70). ■

∞∞,,NM Now consider the functional Z :h ={}CCcc×→ given by

1 d ZY()uv,;t ≡∇⎡⎤ () uv , − idt⎣⎦⎢⎥ut() xy (16) KL 1 −∇∇kl ∇∇ l k ∑∑{}yxF ()t ,0,0kl+ ut() vt ()Y xy()uv , kl==00 i for every tI∈ . Let us assume that Z is of class V ∞ , then, by Theorem 3.3 it can be expanded to the following form 148 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

∞∞ 1 ()uv,;tt= δδnm ()[][] 0,0; u ,…… , u v , , v (17) ZZ∑∑ uv nm==00nm!! n times m times for every tI∈ . It is clear that Z ()uv,;t = 0 iff

δδnm ()[][]0,0;,,tnmtIuuvv…… ,,=∈∈ 0,for every (), 2 and for every . uvZ + n times m times

Clearly every term δδnm ()[][]0,0;,,t uuvv…… ,, for all ()nm, ∈ 2 of the sum (17) is a uvZ + n times m times homogeneous functional polynomial, i.e. there exists an ()nm, −multilinear mapping

∞∞,,NMnm ^ :()()CCcc××→ I with the symmetry property

^^[][]uuvv,,…… ,,= ⎡ u ,, … uvv⎤⎡ ,, … ⎤ 11nm⎣⎢ iijj11nm⎦⎣⎥⎢ ⎦⎥

for every permutation ()ii1,,… n and ()j1,,… jm , such that,

δδnm ()[][][][]0,0;,,t uuvv…… ,,= uuvv ,, …… ,, uvZ^ n times m times n times m times

For the homogeneous functional polynomials the following important property holds, ∞,N Preposition 3.5 : Let the homogeneous functional polynomial of degree n , Φnc:C →

∞,N n described by the corresponding n−multilinear, symmetric function ^ :()Cc → . Then ∞,N the polynomial Φnc:C → is the zero polynomial iff

⎡⎤ ^ δδ,,,… δ = 0, for every (tt12,,… , tn ) with t ∈×I … ×I . ⎣⎦⎢⎥tt12 tn i Ntimes

∞,N Proof : Clearly, the above assertion for a functional polynomial of degree n , Φnc:C → is ∞ equivalent with the same assertion for a functional polynomial of degree nN⋅ , ΦnN⋅ :C c → . Hence, its sufficient to prove the preposition for the case when N =1. The general form of a homogeneous functional polynomial with N =1 will be of the form

uu,,…=…… sssusususdsds ,,, …… ^ []11211221nnnnn∫∫ ∫ K ( )()()()

If ⎡⎤ ^ δδtt,,,… δ t= 0, for every ()tt,,… , t with tI∈ ⎣⎦⎢⎥12 n 12 n i then

K ()ss12,,,… sn = 0, for every ()ss12,,,… sn with sIi ∈

Hence, Φn will be the zero polynomial.

If Φn is the zero polynomial then using Schwarz’s Inequality we will have

uu,,…=…… sssusususdsds ,,, …… ≤ ^ []11211221nnnnn∫∫ ∫ K ( )()()()

n ⎛⎞⎟ ==⎜ ……K ss,, , s usus …… us dsds⎟ 0 ∏⎜∫∫ ∫ ()()()()12ni 1 i 2 in 1 n⎟ i=1 ⎝⎠⎜ ⎟ 5.3. REDUCTION OF THE HOPF-TYPE FUNCTIONAL DIFFERENTIAL EQUATIONS FOR SPECIAL CASES 149

Hence, we have proved the assertion ■

Thus we have proved the following Theorem 3.6 : Let the Functional Differential Equation

11d KL ⎡⎤∇=∇∇∇∇()uv,,0,0,kl F (t ) l k () uv ⎣⎦⎢⎥ut()YYxy∑∑{} y xkl+ ut() vt () xy idtkl==00 i N Y xy ()()υυ⋅=δ00,0 Y , υ ∈

∞,N ∞,M ∞ for every u ∈CIc (), v ∈CIc () and tI∈ . We assume that Y xy is of class V . Then the above Functional Differential Equation is equivalent with the infinite set of equations,

1 d ⎡⎤δδnm∇= Y ()0, 0⎡⎤δδδδ ,…… , , , , idt⎣⎦⎢⎥uvut() xy⎣⎦⎢⎥ t11 tkk s s KL (18) kl1 nml k ⎡ ⎤ =∇∇y xuvxF ()t,0,0δδ ∇∇Y y () 0,0δδδδttss ,…… , , , , ∑∑{}kl+ ut() vt () ⎣⎢ 11nm⎦⎥ kl==00 i

2 N M for all ()nm, ∈ + and every t , (tt12,,… , tk ) and ()ss12,,,… sl with tI∈ , ti ∈ I , si ∈ I . The above set of equations will be called the corresponding moment equations of problem Σ1. ■

Remark 3.7 : For the case when the first (kl, ) moment equations satisfied, we can get, using Theorem 3.3, useful bounds for the error of the estimation. ■

Remark 3.8 : The above Theorem presents a valuable method for solving numerically functional differential equations of the form (7)-(8). ■

Example 3.9 : Let us now apply the above Theorem to the specific case of the Duffing Oscillator. As we saw, the corresponding functional differential equation is

md2 δδδδYYYY()uv,,,, cd() uv k () uv1 3 () uv xy++⋅+⋅= xy xya xy idt2 δδδ ut() idt ut() i ut() i3 δut()3

1 δY xy ()uv, =⋅ ivtδ () Y xy ()δ00,0 = Y and Y xy ()δ00′,0 = Y

∞ ∞ for every u ∈CIc (),v ∈CIc () and tI∈ . Applying Theorem 3.6 we will have the equivalent set of moment equations,

dd2 δδnm++11YY()0,0 nm ++ () 0,0 mcδδnm xy++ xy dt2 uvδδ u u…… δ u δ v δ u dt δδ u u …… δ u δ v δ u tt11 tnn t++ t nm tt 11 t nn t ++ t nm nm++13 nm ++ δδYYxy()0,0 xy () 0,0 +−ka = (19) δδuu…… δ u δ v δ u δ u3 δ u …… δ u δ v δ u tt11 tn t n++ t nm t t 11 t n t n ++ t nm δ nm++1Y ()uv, = xy δδvu… δ u δ v …δu tt1 tn tnnm++1 t for all ()nm, ∈ 2 and every ()tt,,, t… , t with tI∈ , where []δ = ⎡δδδδ,,,,,……⎤ . ■ + 12 nm+ i ⎣⎢ ttss11kk⎦⎥ 150 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

5.3.3. Excitation with Independent Increments - Fokker-Planck Equation

In this section we will derive from the Functional Differential Equation the Fokker-Planck type equation for the characteristic function of the response. It is very important to note that the analysis below will be based only on the concept of Functional Differential Equations. Moreover the Fokker-Planck Equation derived can describe systems under any stochastic excitation with independent increments. To begin the derivation of the Fokker-Planck equation let us first consider the special case of a first order equation with cubic nonlinearity

3 dxtt() t;;ββ= axt () dt+ dy () t ; β a.s. (20)

xt()()00;ββ= x , a.s. (21) where yI()ii;:×Ω → is a stochastic process with independent increments defined onto the

∞∞ probability space ()ΩΩ,,Uc() with the induced probability space ()CCcc,,Uc()y , and with characteristic functional Y y . The initial condition x0 ()i:Ω→ is described by a random variable with the induced probability space ,,() with characteristic ()Ucx0 function Y 0 . By Theorem 2.4 we have the equivalent Functional Differential Equation

⎡⎤δδYYY()uv,,3 () uv ⎡⎤ δ() uv , dadtd⎢⎥xy=− xy + ⎢⎥ xy (22) tt⎢⎥3 ⎢⎥ ⎣⎦⎢⎥δδut() δut() ⎣⎦⎢⎥ vt()

YYxy ()υδ00,0=∈ () υ , υ (23)

Setting to the above FDE ust=⋅υδ()i −− ,v = 0 at time ts+ , and taking the limit s → 0 we have for the l.h.s. term,

⎛⎞ δY xy ()υδ⋅−−()i st,0 ⎜ ⎟ ⎡⎤ limdssx⎟=++= lim idxts() exp{} ixtsυ ()c () dx ss→→00⎜ ⎟ ∫ ⎣⎦ ⎜ δut()+ s ⎟ ∞ ⎝⎠Cc ⎡⎤ 111⎢⎥ ∂φυx (),t =+=+=limdixtsdxdtsdssxsx⎢⎥ exp{}υφυ()c () lim() , ss→→00∫ υυυ⎢⎥∞ ∂t ⎣⎦Cc

For the first r.h.s. term we have,

3 3 δυδ()i−−st,0 3 ∂ φυ,t Y xy ()⎡⎤ x () limdt=+ lim ixts() exp{} ixtsυ () +=c x () dx ss→→003 ∫ ⎣⎦ 3 δut()+ s ∞ ∂υ Cc

Finally for the second r.h.s. term we have

⎡⎤δυδ⋅−−i st,0 ⎢⎥Y xy ()() limdssxy=++ lim idyts() exp{} ixtsυ ()c () dxdy , ss→→00⎢⎥∫∫ ⎢⎥δvt()+ s ∞∞ ⎣⎦CCcc

But exp()ixtυυυυ()+= s exp() ixt() + iaxt ()3 s ⋅ exp() i() yt()() +− s yt Hence, 5.3. REDUCTION OF THE HOPF-TYPE FUNCTIONAL DIFFERENTIAL EQUATIONS FOR SPECIAL CASES 151

limidsxy y() t++ s exp{} iυ x () t sc () dx , dy = s→0 ∫∫ CC∞∞ cc 3 =++⋅+−limidyts ( s ) exp iυυ xt () iaxts () exp i υ() yt()() s ytc xy () dxdy , s→0 ∫∫ ()() ∞∞ CCcc

Since the stochastic process yI()ii;:×Ω → has independent increments, and the quantity exp()ixtυυ()+ iaxt ()3 s depends only to prior increments, we have,

3 limidyts (++⋅+−= s ) exp iυυ xt () iaxts () exp i υ() yt()() s ytc xy () dxdy , s→0 ∫∫ ()() ∞∞ CCcc

=+limidytsy() s exp iυ() yt ()() +−⋅ s ytc () dy s→0 ∫ () C∞ c 3 ⋅+lim exp iυυ x() t i ax () t sc x () dx = s→0 ∫ () ∞ Cc 1 =+−⋅=limdiytsytdyixtdxsyx expυυ()()()cc() exp() () () s→0 ∫∫() υ ∞ ∞ CCc c ⎡⎤ 1 ⎢⎥ =+−⋅=limdiytsytdyixdxsyx⎢⎥ expυυξ()()()cc() exp{} () () s→0 ∫∫() υ ⎢⎥∞ ∞ ⎣⎦CCc c

1 ∂φυΔy ()ts,; =⋅limφυx ()tds , υ s→0 ∂s

Thus we have proved the equality

3 11∂∂φυxx(),,tt φυ() ∂φυΔy ()ts,; ds=− a ds + limφυx () t , ds υυυ∂∂ts3 s→0 ∂ or equivalently,

3 ∂∂φυxx(),,tt φυ() ∂φυΔy ()ts,; +=atυφυlimx () , ∂∂tsυ3 s→0 ∂

Assume now, that yI()ii;:×Ω → is the standard Wiener process. Then,

⎪⎪⎧⎫1 2 φυΔy ()ts,;=− exp⎨⎬ υ s ⎩⎭⎪⎪2 Hence, ∂φυ()ts,; 1 lim Δy =− υ2 s→0 ∂s 2 and the Fokker-Planck equation takes the form

∂∂φυ(),,tt3 φυ() υ2 xx+=−atυφυ(), ∂∂t υ3 2 x For the general case we have the following

152 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

Theorem 3.10 : Let the dynamical system Σ2 described by (31)-(32) of Section 5.2.4. Then the FDE

L 11⎡⎤ll dtdtt ∇=∇∇YYxy()uv,,0,{} x F () xy () uv + ii⎣⎦⎢⎥ut() ∑ l ut() l=0 (24) K 1 +∇kkGuv()td,0 ∇⎡ ∇ () , ⎤ ∑{}xxyk+1 ut()⎣⎢ t vt ()Y ⎦⎥ k=0 i

Y ()υ⋅=δ00,0 Y (25)

∞,N ∞,M for every u ∈CIc (),v ∈CIc () and tI∈ , which describes the system, can be reduced to the partial differential equation

∂φ ()t,υ NLυ x =∇∇+n ll φ υ ∑∑ l+1 {}xxFtn (),0{}υ () t , ∂tinl==10 ⎛⎞N ⎜ ⎟ (26) ∂φυs ⎜tGtt,,;()x() s⎟ Δ y ⎜ ∑ ννm ⎟ ⎧⎫N ⎝⎠ν=1 ⎪⎪ + lim exp⎨⎬ixtνν()υ c x ( dx ) ∫ s→0 ⎪⎪∑ ∞,N ∂s ⎪⎪ν=1 Cc ⎩⎭ where φ()ii,:I ×→ N is the characteristic function of the system response x()t ;β and φ ii,:I ×→ M is the characteristic function of the independent increments Δs y ()

Δ=+−s yy()ttst;β ( ) y (). ■

Proof : We consider the component wise form of the FDE (24), used for the derivation at Theorem 2.4. (nN= 1,2,… , )

11δδYY()uv,,L ∂l Ft(),0 l () uv xy=⋅+n xy ddtt ∑∑l iutδδn () l=0 ii∈I N ∂xuii() t i =l (27) MK k k ⎡ ⎤ 1 ∂ Gt(),0 δ δY xy ()uv, + nm ⎢d ⎥ ∑∑k+1 ∑ ⎢ t ⎥ itvtN ∂xuδδ mk==10 i∈I ii()⎣⎢ m () ⎦⎥ i =k

Setting to the system of functional differential equations u =⋅−−()()υυ12,,…i υN δ ts at time ts+ and t waking the limit s → 0 we have for the l.h.s. term,

υ⋅−−δ i ts,0 ⎧⎫N 11δY xy ()() ⎪⎪ limdtnsn=++ lim idxtsυυ() exp⎨⎬ i xtsνν ()c x () dx ss→→00∫ ⎪⎪∑ iutsiδυnn()+ ∞,N ⎪⎪ν=1 Cc ⎩⎭

For the first r.h.s term we have,

L l l 1 ∂+Ft() s,0 δ Y xy ()υ⋅−−δ()i ts,0 lim n ⋅= s→0 ∑∑l l=0 itsi∈I N ∂+xuiiδ () i =l

LNl ∂+Ft s,0 i ⎧⎫ 1 n ()l ⎡⎤⎪⎪ =⋅++=limitsxx() exp⎨⎬ ixtsνν ()υ c x () d s→0 ∑∑l ∫ ⎣⎦ ∑ i N ∂x ⎪⎪ l==01i∈I i C∞,N ⎩⎭⎪⎪ν i =l c 5.3. REDUCTION OF THE HOPF-TYPE FUNCTIONAL DIFFERENTIAL EQUATIONS FOR SPECIAL CASES 153

L ll 1 iυn ∂∂Ftn (),0φx () t ,υ =⋅∑∑l . iiυn l=0 i∈I N ∂∂xiiυ i =l Where we have used the notation

ii12 iN ∂=∂⋅∂xi x12xx… ⋅∂N

i ii12 iN x =⋅⋅⋅x12xx… N

For the second r.h.s term we have

MK k k ⎡ δ υ⋅−−δ i ts,0 ⎤ 1 ∂+Gtsnm (),0 δ ⎢ Y xy ()()⎥ lim ⋅=ds s→0 ∑∑k+1 ∑ ⎢ ⎥ mk==10itsvtsi∈I N ∂+xuiiδδ()⎢ m () +⎥ i =k ⎣ ⎦

MK k 1 ∂+Gts(),0 i =+⋅lim nm itsk+1 ⎡⎤x() s→0 ∑∑k+1 ∑ ∫ ⎣⎦ i N ∂x mk==10 i∈I i CC∞∞,,NM× i =k cc ⎪⎪⎧⎫N ⋅+d y() t s;expβυ⎨⎬⎪⎪ i x ()() t + s dx , dy sm ⎪⎪∑ ννc xy ⎩⎭⎪⎪ν=1

Since x(t) is m.s. continuous lim xx()()ts+= t. Then s→0

MK k 1 ∂+Gts(),0 i lim nm itsk+1 ⎡⎤x()+⋅ s→0 ∑∑k+1 ∑ ∫ ⎣⎦ i N ∂x mk==10 i∈I i CC∞∞,,NM× i =k cc ⎪⎪⎧⎫N ⋅+dyts()exp⎨⎬⎪⎪ i xts () +υ ( dxdy , ) = sm ⎪⎪∑ ννc xy ⎩⎭⎪⎪ν=1 MK k ∂ Gt,0 i 1 nm () k+1 ⎡⎤ =+⋅lim dismx() t yt ( s ) s→0 ∑∑k+1 ∑ ∫ {}⎣⎦ i N ∂x mk==10 i∈I i CC∞∞,,NM× i =k cc ⎪⎪⎧⎫N ⋅+exp⎨⎬⎪⎪i x() t sυ ( dx , dy ) = ⎪⎪∑ ννc xy ⎩⎭⎪⎪ν=1 Using the equation of the system, or the equation that results from the FDE by setting u = 0 , v = 0 we will have for small s M +=ββ + β + β⎡ +− ββ⎤ xtnnn( s; ) xt () ; Ft() ,;xx () t s∑ G nmmm() t ,;() t⎣ y()() t s ; y t ;⎦ m=1 Hence the last term above will be equal with MK k N ∂ Gt,0 i ⎧⎫ 1 nm () k+1 ⎡⎤ ⎪⎪ =+limdismx() t yt ( s ) exp⎨⎬ i xtνν()υ s→0 ∑∑k+1 ∑∫ {}⎣⎦ ∑ i N ∂x ⎪⎪ mk==10 i∈I i CC∞∞,,NM× ⎩⎭⎪⎪ν=1 i =k cc ⎧⎫NM ⎪⎪⎛⎞ ⋅++−=exp⎨⎬iFttsGttytsyt⎜ () ,xx() () ,()⎡⎤()()⎟υ ( dxdy , ) ⎪⎪∑∑⎜ ννmm⎣⎦ m⎟ νc xy ⎩⎭⎪⎪ν==11⎝⎠m 154 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

MK k N ∂ Gt,0 i ⎧⎫ 1 nm () k+1 ⎡⎤ ⎪⎪ =+limdismx() t yt ( s ) exp⎨⎬ i xtνν()υ s→0 ∫ ∑∑k+1 ∑{}⎣⎦ ∑ i N ∂x ⎪⎪ CC∞∞,,NM× mk==10 i∈I i ⎩⎭⎪⎪ν=1 cc i =k ⎪⎪⎧⎫NM ⋅+−exp⎨⎬⎪⎪iGttytsytdxdy() ,x()⎡⎤() ()υ ( , ) ⎪⎪∑∑ ννmmm⎣⎦c xy ⎩⎭⎪⎪ν==11m

Thus for every nN=1,2,… , equation (27) takes the form

⎧⎫NL∂∂llFt,0φ t ,υ ⎪⎪iυn n () () limidxtυυnsn()++= s exp⎨⎬ i xtνν () sc x () dx ⋅+ ds s→0 ∫ ∑∑∑l ⎪⎪i N ∂∂x υ C∞,N ⎩⎭⎪⎪ν==10l i∈I ii c i =l

MK k N υ ∂ Gt(),0 i ⎪⎪⎧⎫ n nm k+1 ⎡⎤ ⎪⎪ ++limdismx() t yt ( s ) exp⎨⎬ i xtνν()υ s→0 ∫ ∑∑k ∑{}⎣⎦ ∑ i N ∂x ⎪⎪ CC∞∞,,NM× mk==10 i∈I i ⎩⎭⎪⎪ν=1 cc i =k ⎪⎪⎧⎫NM ⋅+−exp⎨⎬⎪⎪iGttytsytdxdy() ,x()⎡⎤() ()υ ( , ) ⎪⎪∑∑ ννmmm⎣⎦c xy ⎩⎭⎪⎪ν==11m

Summing up for all nN=1,2,… , , we will have for the l.h.s. sum

NN ⎪⎪⎧⎫ limidxtυυnsn()++= s exp⎨⎬ i xtνν () sc x () dx s→0 ∑∑∫ ⎪⎪ n==11∞,N ⎪⎪ν Cc ⎩⎭

⎡⎤N ⎢⎥⎪⎪⎧⎫ =+=limdixtsds ⎢⎥ exp⎨⎬νν()υ c x ()x s→0 ∫ ⎪⎪∑ ⎢⎥∞,N ⎩⎭⎪⎪ν=1 ⎣⎦Cc

∂φx ()t,υ =+=limdtssφx () ,υ ds s→0 ∂t

For the first r.h.s sum we get,

NLiiυυ∂∂llFt(),0φ () t ,υ NL nnn ⋅=∇∇x ll φ υ ∑∑ll ∑ds ∑∑ {}xx Fn () t,0{}υ () t , nl==10iii∈I N ∂∂xiiυ nl == 10 i =l

For the second r.h.s sum we have,

NMK k υ ∂ Gt(),0 i n nm k+1 ⎡⎤ lim dismx() t yt (+⋅ s ) s→0 ∫ ∑∑∑k ∑ {}⎣⎦ i N ∂x CC∞∞,,NM× nmk===110 i∈I i cc i =k ⎪⎪⎪⎧⎫⎧NNM ⎪⎫ exp⎨⎬⎨⎪⎪⎪i x() tυυ exp i G() t ,x () t⎡⎤ y() t+− s y () t ⎬⎪ ( dx , dy ) = ⎪⎪⎪∑∑∑νν νmmm⎣⎦ ν ⎪c xy ⎩⎭⎩⎪⎪⎪νν===111m ⎭⎪ ⎡⎤⎧⎫NM⎪⎪ N ⎢⎥⎪⎪ =⋅+⋅limdiGttytsixtsnmmυυννν() ,x() () exp⎨⎬ () s→0 ∫ ⎢⎥∑∑⎪⎪ ∑ ∞∞,,NM⎣⎦⎩⎭nm==11⎪⎪ν = 1 CCcc× ⎪⎪⎧⎫NM ⋅+−=exp⎨⎬⎪⎪iGttytsytdxdy() ,x()⎡⎤() ()υ ( , ) ⎪⎪∑∑ ννmmm⎣⎦c xy ⎩⎭⎪⎪ν==11m 5.3. REDUCTION OF THE HOPF-TYPE FUNCTIONAL DIFFERENTIAL EQUATIONS FOR SPECIAL CASES 155

⎧⎫MN ⎪⎪⎡⎤ =+−⋅limdiGttytsytsmmm exp⎨⎬νν() , x() () ()υ s→0 ∫ ⎪⎪∑∑ ⎣⎦ ∞∞,,NM ⎩⎭⎪⎪m==11ν CCcc× ⎪⎪⎧⎫N ⋅exp⎨⎬⎪⎪ixt()υ ( dxdy , ) ⎪⎪∑ ννc xy ⎩⎭⎪⎪ν=1 Since the stochastic process y()ii;:I ×Ω → M has independent increments, and the quantity ⎪⎪⎧⎫N ⋅exp⎨⎬⎪⎪ixt()υ depends only to prior increments, the last expression will be equal with ⎪⎪∑ νν ⎩⎭⎪⎪ν=1 ⎧⎫MN ⎪⎪⎡⎤ limdiGttytsytdysmmm exp⎨⎬νν() , x() ()+− ()()υ c y ⋅ s→0 ∫ ⎪⎪∑∑ ⎣⎦ ∞∞,,NM ⎩⎭⎪⎪m==11ν CCcc× ⎪⎪⎧⎫N ⋅=exp⎨⎬⎪⎪ixt()υ ( dx ) ⎪⎪∑ ννc x ⎩⎭⎪⎪ν=1 ⎛⎞⎧⎫NN ⎜ ⎟ ⎪⎪ ==limdtGttsixtdxsmφυυs ⎜ ,νν() ,x() ;⎟ exp⎨⎬ νν()c x ( ) s→0 ∫ Δ y ⎜ ∑∑⎟ ⎪⎪ ∞,N ⎝⎠νν==11⎪⎪ Cc ⎩⎭ ⎛⎞N ⎜ ⎟ ∂φυs ⎜tGtt,,;()x() s⎟ Δ y ⎜ ∑ ννm ⎟ ⎧⎫N ⎝⎠ν=1 ⎪⎪ = dslim exp⎨⎬ i xνν() tυ c x ( dx ) ∫ s→0 ⎪⎪∑ ∞,N ∂s ⎪⎪ν=1 Cc ⎩⎭ As a result we have the desired equation for the characteristic function

∂φ ()t,υ NLυ x =∇∇+n ll φ υ ∑∑ l+1 {}xxFtn (),0{}υ () t , ∂tinl==10 ⎛⎞N ⎜ ⎟ ∂φυs ⎜tGtt,,;()x() s⎟ Δ y ⎜ ∑ ννm ⎟ ⎧⎫N ⎝⎠ν=1 ⎪⎪ + lim exp⎨⎬ixtνν()υ c x ( dx ) ■ ∫ s→0 ⎪⎪∑ ∞,N ∂s ⎪⎪ν=1 Cc ⎩⎭

Remark 3.11 : We must note that the essential property for the proof of the Fokker-Planck equation is the independent increment property of the excitation, since we can decouple the probabilistic structures of the excitation and the response at the last term of the functional equation. Hence we can write equations for the response without taking into account the correlation structure of the two stochastic processes. This is not the case, when the stochastic excitation does not posses the independent increment property. In this case we must consider the correlation structure of the problem.

Remark 3.12 : As we mention at the beginning the above proof of the Fokker-Plank equation is based only on the

Remark 3.13 : The limit procedure with respect to s is carried out to take into account the discontinuity of the excitation.

Corollary 3.14 : Let the dynamical system Σ2 described by (31)-(32) of Section 5.2.4, with the excitation process being the standard Wiener process W ()t ;β tI∈ with intensity vector dW (t),tI∈ , i.e., with characteristic function of increments

156 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

⎪⎪⎧⎫⎛⎞M ⎪⎪1 ⎜ 2 ⎟ φ s ()ty;,,… y=− exp⎨⎬⎜ yd() t⎟ s Δ y 1,MiWi⎪⎪⎜∑ ⎟ ⎩⎭⎪⎪2 ⎝⎠i=1 then eq. (26) takes the form

∂φ ()t,υ NLυ x =∇∇+n ll φ υ ∑∑ l+1 {}xxFtn (),0{}υ () t , ∂tinl==10 kk 1 NMK ∂∂12Gt(),0 Gt() ,0 ∂kk12+ φ ()t,υ (28) −⋅⋅υυ dt() νν21mm x ∑∑∑∑νν12 Wm, 2 N ∂∂∂xxυ ν12===110mk ij, ∈I iji+j ν == 2110k i =k1 j =k2

N where φx ()ii,:I ×→ is the characteristic function of the system response x()t ;β . ■

Proof : By direct calculation we have

⎛⎞N ⎜ ⎟ ∂φυs ⎜tGtt,,;()x() s⎟ Δ y ⎜ ∑ ννm ⎟ ⎧⎫N ⎝⎠ν=1 ⎪⎪ lim exp⎨⎬ixtνν()υ c x ( dx )= ∫ s→0 ⎪⎪∑ ∞,N ∂s ⎪⎪ν=1 Cc ⎩⎭

MN2 N 1 ⎛⎞⎧⎫⎪⎪ =−dt()⎜ Gtt(),expxx ()υυ⎟ ⎨⎬ ixt() ( d ) = ∫ ∑∑Wm, ⎜ νν m ⎟ ⎪⎪ ∑ ννc x ∞,N 2 m==11⎝⎠νν⎩⎭⎪⎪ = 1 Cc

MN 2 N 1 ⎛⎞⎪⎪⎧⎫ =− dt()⎜ Gtt(), x ()υ ⎟ exp⎨⎬ixt()υ ( dx )= ∑∑Wm, ∫ ⎜ νν m ⎟ ⎪⎪∑ ννc x 2 m==11∞,N ⎝⎠ν ⎪⎪ν=1 Cc ⎩⎭

MN N 1 ⎪⎪⎧⎫ =−dtWm, ()υυνν GttGtt ν m(),,expxx () ν m()() ⎨⎬ ixt ν() υ νc x ( d x ) = 2 ∑∑12∫ 2 1 ⎪⎪ ∑ m==11νν1 ∞,N ⎪⎪ = 1 Cc ⎩⎭ ν2 =1 ⎧⎫N ⎪⎪ NM ⎨⎬⎪⎪ixt()υ ⎪⎪∑ νν 1 ⎩⎭⎪⎪ν=1 =− υυννdtGttGtteWm, () ν m(),,xx () ν m()() c x () d x (29) 2 ∑∑12∫ 2 1 ν1==11∞,N m Cc ν2 =1

Using the the component wise form the Taylor expansion of Gtνm (), x() twe have

MK∂∂kk21Gt(),0 K Gt() ,0 ⎪⎪⎧⎫ N dtνν12mmxxjiexp⎪⎪ ixtυ d x= ∫ ∑∑∑Wm, () ∑∑⎨⎬ ∑νν()c x ( ) NN∂∂xx⎪⎪ ∞,N mk==1021ji∈∈IIji k = 0⎪⎪ν = 1 Cc ⎩⎭ ji==kk21 kk MK∂∂21Gt(),0kk21 K Gt() ,0 ⎪⎪⎧⎫N νν12mm∂∂⎪⎪ = dtWm, () exp⎨⎬ ixν ()tdυν c x (x )= ∑∑∑kk21 ∑∑ ∫ ∑ NNii∂∂xxυυ ∂∂ ⎪⎪ mk==1021ji∈∈IIjj k = 0 ii∞,N ⎪⎪ν=1 Cc ⎩⎭ ji==kk21 kk + MK ∂∂12Gt(),0 Gt() ,0 1 ∂kk12φ ()t,υ = dt() νν21mm x ∑∑ ∑ Wm, kk12+ N ∂∂xxi ∂υ mk==102 ij, ∈I ij i+j = k1 0 i =k1 j =k2

Hence we have proved the desired equation. ■

5.3. REDUCTION OF THE HOPF-TYPE FUNCTIONAL DIFFERENTIAL EQUATIONS FOR SPECIAL CASES 157

Remark 3.15 : Taking the Fourier transform of the equation (28) we clearly have for the l.h.s. and r.h.s. term that

⎡⎤∂∂φ ()tft,,υ ()x −1 ⎢⎥xx= F ⎢⎥ ⎣⎦∂∂tt

⎡⎤NLυ −1 ⎢⎥n ll F {}∇∇=xxFtn (),0{}υφ () t ,υ ⎢⎥∑∑il+1 ⎣⎦nl==10 N ∂ =−⎡⎤ ∈ ∈ N ∑ ⎣⎦ftx ()(),,,,xx Ftn t I x n=1 ∂xn

For the second r.h.s. term we have from (29)

⎡⎤⎛⎞N ⎢⎥⎜ ⎟ ∂φυs ⎜tGtt,,;()x() s⎟ ⎢⎥Δ y ⎜ ∑ ννm ⎟ ⎧⎫N −1 ⎢⎥⎝⎠ν=1 ⎪⎪ F lim exp⎨⎬ixt∑ νν()υ c x ( dx )= ⎢⎥∫ s→0 ∂s ⎪⎪ ⎢⎥C∞,N ⎩⎭⎪⎪ν=1 ⎢⎥c ⎣⎦ ⎧⎫N ⎡⎤⎪⎪ NM ⎨⎬⎪⎪ixt()υ 1 ⎢⎥⎪⎪∑ νν −1 ⎢⎥⎩⎭⎪⎪ν=1 =−F υυννdtGttGtte() ν(),,xx () ν()() c () d x = ⎢⎥∑∑12∫ Wm, 2 m 1 m x 2 ν ==112,N m ⎢⎥1 _ ⎣⎦ν2 =1 N 1 ∂∂ T = ⎡⎤Gxd()tttft,,, ()()() Gx x 2 ∑ ∂∂xx⎣⎦⎢⎥Wm, x ν1=1 νν12 ν2 =1

⎛⎞ ⎜dtW ,1 () … 0 ⎟ ⎜ ⎟ where d ()t = ⎜ ⎟ is the intensity matrix of the Wiener process W ()t ;β Wm, ⎜ ⎟ ⎜ ⎟ ⎝⎠⎜ 0 dtWM, ()⎟ tI∈ . Hence we have proved the Fokker-Planck equation for the probability density function

NN ∂ft(), x ∂∂∂1 T x +−⎡⎤ft()(),,xx Ft⎡⎤ GxdGxx() t , ()()() t t , ft ,0 = ∂∂tx∑∑⎣⎦x nWm2 ∂∂ xx⎣⎦⎢⎥, x n==11n ν1 νν12 ν2 =1 for ()tI, x ∈× N . ■

Example 3.16 [One-dimensional case] : Let us now consider a specific case of a one dimensional system excited with from a Wiener stochastic process W ()t ;β tI∈ with constant intensity dW and coefficients independent of time. The Fokker-Planck equation takes the form

2 ∂ftx(), ∂∂1 2 x +−⎡⎤ftxFx()(),,0⎡⎤ dGxftx() ( ) = (30) ∂∂tx⎣⎦xx2 ∂ x2 ⎢⎥⎣ W ⎦

Using the standard method of separation of variables we assume that

fx ()tx, =Λ ()() xTt (31) where Λ and T are functions of x and t , respectively. Substitution of (31) into (30) gives the two ordinary differential equation 158 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

1 dT =−λ2 (32) Tt() dt 2 1 ∂∂2 ⎡⎤dGx()Λ− () x⎡⎤ Fx() Λ+Λ= () xλ2 () x 0 (33) 2 ∂∂xx2 ⎢⎥⎣⎦W ⎣⎦

Equation (32) is a simple first order equation with solution

Tt()=−exp{}λ2 t (34) while equation (33) is a second order equation with variable coefficients. In general, the solution Λ()x,,,ΑΒλ depends on two arbitrary constants Α and Β . If the eigenvalues of equation (33) are discrete and distinct, then by virus of linearity of the problem the solution can be represented symbolically in the form

∞ 2 f x ()tx,,,,exp=Λ∑ ( xΑΒλnnn ) {} − λ n t (35) n=0

where ΑΒnn, and λn are determined from the initial and boundary conditions.

In general case of variable coefficients Gx( )2 and Fx() the eigenfunction-eigenvalue problem (33) is not easy to solve. Only in restricted numbers of cases (special form of Gx( )2 and F ()x ) a solution can be determined in a closed form – in terms of special functions. The solution of the Fokker-Planck equation can be then expresses in terms of the eigenfunctions Λn ()x .

For example, if equation (33) is supplemented by the absorbing boundary condition

ftx (),0±= a and the initial condition is f x ()(tx00, =−δ t t ) then the solution is represented as

∞ ΛΛnn()xx (0 ) 2 f x ()tx,exp=−−∑ {}λn () t t0 (36) n=0 fxx,st ()

where f x,st (x) is the stationary solution of equation (30) (which will determined in the sequel) −1 and the functions Λ are orthonormal with the weight function ⎡ f x ⎤ , i.e., n ()x ⎣ x,st ()⎦

a 1 ⎧⎪1, nm= ΛΛ()xxdx () =⎨⎪ ∫ nm ⎪0, nm≠ −a fxx,st () ⎩⎪

The above form of solution is often used in applications.

Now let us determine the stationary solution probability density function for the one dimensional case described above. Note that for the existence of stationary solutions, certain conditions must be fulfilled for coefficients Gx( )2 and Fx(). These conditions are described extensively for the one dimensional case, as well as, for the general case at SOBCZYK, K., (Stochastic Differential Equations. p.154). Now consider the one-dimensional case. Assuming that a stationary solution exists, then since f x,st()xftx= lim x ( , ), the stationary solution does t→∞ not depend on time and the initial density f0 ()x . The Fokker-Planck equation takes the form

5.3. REDUCTION OF THE HOPF-TYPE FUNCTIONAL DIFFERENTIAL EQUATIONS FOR SPECIAL CASES 159

2 ∂∂1 2 ⎡⎤fxFx() ()−=⎡⎤ dGxfx() () 0 (37) ∂∂xx⎣⎦xst,,2 2 ⎣⎦⎢⎥W xst or, after integration with respect to x

d 2 ⎡⎤dGx() f () x−=−22⎡⎤ f () xFx () Φ (38) dx ⎣⎦⎢⎥W xst,,⎣⎦ xst 0

where Φ0 is a constant. The last equation is first order inhomogeneous equation for function 2 hx()= dGxW () fxst, () x,

dh() x F() x −=−22hx() Φ (39) dx 2 0 dGxW ()

Hence, the general solution for f xst, ()x is

⎪⎪⎧⎫xx ⎪⎪⎧⎫y C ⎪⎪22FF()ξξ2Φ0 ⎪⎪() f xst, ()xd=−exp⎨⎬ξξ exp ⎨⎬ ddy (40) dGx222⎪⎪∫∫∫ dGξξ dGx ⎪⎪ dG 2 WWW() ⎩⎭⎪⎪xxx′′′() () ⎩⎭⎪⎪ W()

where the constant C is determined by the normalization condition, whereas the constant Φ0 is found from the boundary conditions; as lower integration limit x′ one can take an arbitrary point from the state space of the process.

If Φ0 = 0 equation (37) gives

d 2 ⎡⎤dGx() f () x−=−22⎡⎤ f () xFx () Φ (41) dx ⎣⎦⎢⎥W xst,,⎣⎦ xst 0 with the following general solution

⎪⎪⎧⎫x C ⎪⎪2F ()ξ f xst, ()xd= exp⎨⎬ξ (42) 22⎪⎪∫ ξ dGxWW() ⎩⎭⎪⎪x′ dG()

The above formula is often used in applications. More general formulas describing the probability density function of Hamiltonian Dynamical Systems with Wiener excitation and/or white noise parametric uncertainty are of special interest. For results of this type, as well as the necessary conditions for their existence we refer to SOIZE, C., (The Fokker-Planck Equation and its Explicit Steady State Solutions) and to POLIDORI, D.C. & BECK, J.L. & PAPADIMITRIOU, C., (A New Stationary PDF Approximation for Non-Linear Oscillators), MOSBAH, H. & FOGLI, M., (An Original Approximate Method for Estimating the Invariant Probability Distribution of a Large Class of Multi-Dimensional Nonlinear Stochastic Oscillators). For other interesting results concerning systems with random excitation of special type we refer to DI PAOLA, M. & SOFI, A., (Approximate Solution of the Fokker-Planck-Kolmogorov Equation) and GUO-KANG ER., (Exponential Closure Method for some Randomly Excited Non-Linear Systems).

160 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

5.3.4. Partial Differential Equation for (N+M)-dimensional joint (excitation-response) characteristic functions in the case of general m.s.-Continuous Excitation

In this section we shall derive a partial differential equation for the joint characteristic function of the excitation and the response when the excitation is m.s.-continuous stochastic process. In contrast with the previous case of an independent increment process where the randomness of the excitation ‘regenerates’ every instant time, in this case the stochasticity evolves in a smoother way, as a result of the finite correlation time of the excitation. This means that, ⎛⎞t ⎜ ⎟ although the response x()t ;β on the past history of the excitation ys⎜ ;β⎟ (Principle of ⎝⎠⎜t0 Causality), the correlation between x()t ;β and yt()+ εβ; , with ε > 0 may not be zero, since

Ctyy ()+≠ε,0 t . Because of the last property of the excitation process, i.e. finite correlation time we cannot use the same arguments as for the previous case of independent increment process. Our analysis will be a generalization of Liouville equation. We shall use the same style of presentation as before, i.e., first a special case will be considered and then the general problem will be addressed. Let us consider a first order equation with cubic nonlinearity

3 x()taxtyt;;;βββ= ()+ () a.s.tt≥ 0 (43)

xt()()00;ββ= x , a.s. (44) where yI()ii;:×Ω → is a m.s. continuous stochastic process defined onto the probability

∞∞ space ()ΩΩ,,Uc() with the induced probability space ()CCcc,,Uc()y , and characteristic functional = expiyv , dy. The initial condition x i:Ω→ is Ycy ∫ {}y () 0 () ∞ Cc described by a random variable with the induced probability space ,,() and ()Ucx0 characteristic function Y 0 (a) , a ∈ . By Theorem 2.1 we know that the joint characteristic functional satisfies

d ⎡⎤δδδYYY()uv,,,3 () uv() uv ⎢⎥xy=−a xy + xy , uv , ∈ C∞ , tt≥ (45) ⎢⎥ 3 c 0 dt⎣⎦⎢⎥δδ u() tδut() v() t

YY()()aaa⋅=δ00,0 , ∈ (46)

– Derivation of a P.D.E. for the joint characteristic function φυxy (),;tvt ,

We shall now derive a partial differential equation for the characteristic function. For the l.h.s. term of (45) we clearly have

⎛⎞ ddδY xy ()uv, ⎟ ⎜ ⎟=+=ix() texp i x , u i y , v() dx , dy ⎜ ⎟ ∫∫ {}c xy dt⎝⎠⎜ δ u() t⎟ dt ∞∞ CCcc× dx() t =+iixuiyvdxdyexp , , , ∫∫ {}c xy () ∞∞ dt CCcc× 5.3. REDUCTION OF THE HOPF-TYPE FUNCTIONAL DIFFERENTIAL EQUATIONS FOR SPECIAL CASES 161

Taking the r.h.s. of the last equality and substituting ut()ii=⋅υδ ( − ) and , vv()ii=⋅δ ( − s ) υ,v ∈ we obtain dx() t iixuiyvdxdyexp{} ,+= ,c () , ∫∫ dt xy CC∞∞× cc dx() t =+=iexp iυ x t ivy s dx , dy ∫∫ {}() ()c xy () ∞∞ dt CCcc× ⎡⎤ 1 d ⎢⎥ =+=⎢⎥exp{}ixtυ () ivys ()c xy () dxdy , υ dt ∫∫ ⎢⎥∞∞× ⎣⎦CCcc 1 d ∞∞ =+exp{}()iυ x ivy f x , t ; y , s dxdy = υ ∫∫ xy dt −∞ −∞ 1 ∂φυ(),;tvs , = xy υ ∂t

For the first r.h.s. term we have,

3 δυδδY ()⋅−()iitv, ⋅− () s 3 =+=⎡⎤ix texp iυ x t ivy s dx , dy 3 ∫∫ ⎣⎦(){} () ()c xy () δut() ∞∞ CCcc× ∂φυ3 (),;tvs , = xy ∂υ3

Finally for the second r.h.s. term we have

δυδY ()⋅−()iitv, ⋅− δ () s =+iy texp iυ x t ivy s dx , dy ∫∫ (){} () ()c xy () δvt() ∞∞ CCcc

Taking the limit st→ and since the excitation has continuous paths we will have

iy texp iυ x t+= ivy t dx , dy ∫∫ (){} () ()c xy () ∞∞ CCcc×

∞∞ ∂φυ(),;tvs , =+iyexp{}() iυ x ivy f x , t ; y , t dxdy =xy ∫∫ xy ∂ −∞ −∞ v

Thus we have proved the following partial differential equation involving the joint characteristic function,

∂∂∂φυ(),;tvs ;3 φυ() ,; tvt , φυ() ,; tvt , xy+=avtTυυ xy xy ,, υ ∈∈ , (47) ∂∂∂tvυ3 st= with the additional condition

φφxy()()0,tvs ; ,==⋅−∈ y vs ,Y y () v δ ()i s , v (48) and the initial condition

φυxy()()(),0;0,s == φυ x ,0Y 0 υ , υ ∈ 162 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

We note that the partial differential equation for the joint characteristic function is of the

Liouville’s equation form with additional condition for the marginal φy ()vs, . In other words, we treat the excitation term yt();β as a state space variable witch doe’s not participate to the dynamical part of the equation. Instead of that, the additional condition (48) is fulfilled.

– Derivation of moment equations for eq. (43) based on eq. (47)

It is worth noticing that equation (47) is equivalent with the infinite set of moment equations

β ⎡⎤⎡⎤nm+1 1 dE()⎣⎦⎣⎦ x() t;;ββ y () s ⋅= ndt+1 (49) st= ββ⎡⎤⎡⎤nm++31 ⎡⎤⎡⎤ nm =+aE()⎣⎦⎣⎦ x() t;;ββ y ( s ) E() ⎣⎦⎣⎦ x() t ;; ββ y () s

for n = 0,1,… and m = 0,1,… . The proof of the above set of moment equation for eq. (43) can be found at BERAN M.J., (Statistical Continuum Mechanics). Let us now derive the above set using directly eq. (47). From Chapter I we clearly have the equation connecting the characteristic function and the moments of a stochastic process. Hence, for the examined case the suitable form of this equation will be

nm+ ∂ φυ(),;tvs , nm xy =⋅iExtnm+ β ⎡ ;;ββ⎤⎡ ys ⎤ , nm,0,1,= … (50) nm ()⎣ ()⎦⎣ () ⎦ ∂∂υ v υ=0 ν=0

Moreover we have,

nm++1 1 ∂ φυ(),;tvs , d nm ⋅=⋅=xy Extβ ⎡⎤⎡⎤;;ββ ys nm+ n m ()⎣⎦⎣⎦() () itvdt∂∂υ ∂ υ=0 ν=0 , nm ,= 0,1,… (51) ⎛⎞ β ⎜dx() t ;β nm−1 ⎟ =⋅nE⎜ ⎡ x() t;;ββ⎤⎡ y ( s ) ⎤⎟ ⎜⎝⎠dt ⎣ ⎦⎣ ⎦⎟

Now, taking n–partial derivatives with respect to υ and m–partial derivatives with respect to ν to eq (47) and using the two last equations (50) and (51) we obtain the set of moment equations (49).

– Derivation of a P.D.E. for the joint density function fxy (xt,; yt ,)

An alternative form of equation (47) can be results by applying the Fourier transform to the above equations. Before that, we recall some useful relations connecting the probability density function and the characteristic function of stochastic processes written in a suitable form for our study. Thus, based on simple properties of the Fourier transform, we have 5.3. REDUCTION OF THE HOPF-TYPE FUNCTIONAL DIFFERENTIAL EQUATIONS FOR SPECIAL CASES 163

∂fxtys(),; , xy =F −1 ⎡⎤iuφ () u,; t v , s ∂x ⎣⎦⎢⎥xy

∂φ ()utvs,; , xy =F ⎡⎤ixf() x,; t y , s ∂u ⎣⎦⎢⎥xy

Based on the above equations we can easily prove the more general

⎡⎤⎡⎤∂∂nnφφ()utvs,; ,1 ∂∂mnm() utvs ,; , i −−11⎢⎥⎢⎥uxfxtysm xy== xy ⎡ n ,; , ⎤ FF⎢⎥⎢⎥nmmnmm⎣⎢ xy ()⎦⎥ ⎣⎦⎣⎦⎢⎥⎢⎥∂∂∂∂uixuix

By direct application of the Fourier transform ant using the last equation we have

∂fxtys(),; , ∂ xy ++=∈∈⎡⎤axfxtytyfxtyt3 ()(),; , ,; , 0, xy , , tI (52) ∂∂tx⎣⎦⎢⎥xy xy st=

∞∞ fxtysdxfys,; ,== ,iyν νδ ⋅−∈i sd ν , y (53) ∫∫xy() y ()e Y y () () −∞ −∞

∞∞ fxysdyfx,0; ,== ,0ixυ υυ d , x ∈ (54) ∫∫xy()() x e Y 0 () −∞ −∞

Thus, the l.h.s. of equation (52), which is equivalent with the l.h.s. of Liouville equation express the conservation of ‘probability’ for the state space variable x()t ;β . The r.h.s term, which is due to the excitation term, express the transfer-interaction of ‘probability’ from the stochastic process y()t ;β to the state space variables x()t ;β . Of course the stochastic characteristics of the excitation y()t ;β are introduced with condition (53).

– Numerical scheme for the solution of eq. (52)

For the case of a first order equation with cubic non-linearity we have proved at the beginning of the section the following partial differential equation involving the joint characteristic function,

∂∂∂φυ(),;tvs ,3 φυ() ,; tvt , φυ() ,; tvt , xy+=avtTυυ xy xy ,, υ ∈∈ , ∂∂∂tvυ3 st= with the additional condition

φφxy()()0,tvs ; ,==⋅−∈ y vs ,Y y () v δ ()i s , v and the initial condition

φυxy()()(),0;0,s == φυ x ,0Y 0 υ , υ ∈

Taking the Fourier transform of the above equations we have the equivalent set of equations for the joint probability density function,

164 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

∂fxtys(),; , ∂ xy ++=∈∈⎡⎤axfxtytyfxtyt3 ()(),; , ,; , 0, xy , , tT (55) ∂∂tx⎣⎦⎢⎥xy xy st= ∞∞ fxtysdxfys,; ,== ,iyν νδ ⋅−∈i sd ν , y (56) ∫∫xy() y ()e Y y () () −∞ −∞ ∞∞ fxysdyfx,0; ,== ,0ixυ υυ d , x ∈ (57) ∫∫xy()() x e Y 0 () −∞ −∞

For the numerical solution of the above problem we will use a combination of a bivariate copula function, to represent the correlation structure of the excitation and the response at time t, and a kernel density function representation for the marginal fx ()xt, . We briefly, recall that a bivariate copula function is a distribution function that has given marginals. For more information we refer to JOE, H., (Multivariate Models and Dependence Concepts). For this example we will use the Plackett copula function that gives a joint density function representation at time t,

fxy()xt,; yt ,=⋅⋅ C() F x ()()() xt , , F y yt , ;ψ t f x() xt , f y () yt , (58) with,

ψψ()tt⋅−⋅+−+()() () 121() xxxx12 12 Cxx()12,;ψψ() t=>3 ,() t 0 (59) 2 2 1141++xxψψψ t − − t t − xx {}()()()12()()() () 12

Additionally we assume that the correlation structure between the response and the excitation, which is described by the function ψ(t), doesn’t change considerable for the joint density function fxy ()xt,; ys , when the time difference ts− is sufficiently small. In this way we can assume that for small ts− the following representation for the joint characteristic function holds

fxy()xt,; ys ,=⋅⋅ C() F x ()()() xt , , F y ys , ;ψ t f x() xt , f y ( ys , ) (60)

Hence, substituting representation (60) into eq. (55) we will have

⎪⎪⎧⎫⎛⎞∂∂CC ∂fxt(), ⎪⎪⎜ ′ ⎟ x ⎨⎬⎜ fxtxx(),,+++ψ () tfxtC⎟ ( ) ⎪⎪⎝⎠⎜∂∂xtψ ⎟ ∂ ⎩⎭⎪⎪1 (61) ∂ ++=∈∈⎡⎤ax3 C f() x,,0,,, t y C f () x t x y t I ∂x ⎣⎦⎢⎥xx

For the last equation the unknowns are the functions fx ()xt, and ψ(t). Now using a kernel density representation for the marginal density function fx ()xt, and multiplying and ingrate over x with each kernel we will have a set of integrodifferential equations for the parameters of the kernel density functions and the function ψ(t). The above approach should be clearer at Section 5.4 where a kernel representation will considered at the level of characteristic functional. An alternative approach can be based on the kernel representation of the correlation structure, i.e., two dimensional kernels that include the correlation structure. Foe more details we refer to Section 5.4. 5.3. REDUCTION OF THE HOPF-TYPE FUNCTIONAL DIFFERENTIAL EQUATIONS FOR SPECIAL CASES 165

– Derivation of a P.D.E. for the joint characteristic function φυυxxyy( 1212,;ttvtvt , ; ,; ,)

For many practical applications it is often need to know the joint probability of the response and its derivative. We shall show that for the case of the joint probability of the response and its derivative, we can easily have the desired density by a simple transformation of random variables x()()tyt;,ββ ; using their joint density derived before. In contrast for higher order derivatives another approach has to be considered. More precisely the joint characteristic function of the response, its derivative and the excitation can be easily derived using the equation governing the system,

3 x()taxtyt;;;βββ= ()+ () a.s.tt≥ 0

For the joint probability of the second derivative we consider the following set of equations

xt();;;βββ=+ axt ()3 yt () a.s.tt≥ 0 (62) x()taxtxtyt;3;ββββ=+ ()()()2 ; ; where the second equations has been derived by direct differentiation of the first one. Now using same arguments as for the case of the derivation of the joint characteristic function

φυxy (),;tvt , we have the equation

∂∂∂φφφ()ts,,,33() tt() tt xxyy++aaυυ xxyy 3 xxyy = ∂∂∂∂t 12υυυ32 st= 112 ∂∂φφxxyy()tt,, xxyy () tt =+υυ12,,,,, υυ 1212vv ∈∈ t T ∂∂vv12 where for simplicity, we write the joint characteristic function φυυxxyy()1212,;ttvsvs , ; , ; , as

φxxyy()ts, . Moreover the additional condition holds

φφδδxxyy(0,t ;0, tvsvs ;12 , ; , )==⋅−+⋅−∈ yy ( vsvs 12 , ; , )Y y () v 1 ()ii s v 2 () s , vv 12 , and the initial condition

φxxyy()()() υ12,0; υ ,0;0,ss ;0,== φ xx υ 12 ,0; υ ,0Y 012 υυ , , υυ 12 , ∈

Similarly we can derive equations for joint characteristic functions of higher derivatives.

– Derivation of a P.D.E. for the joint ch. function φυυxxyy ()12,;tt ,++ τ ; vtvt 12 ,; , τ

For many applications it is often need to know the joint probability of the response for two distinct time instants. More precisely the joint characteristic function of the response at two distinct time instants t and t +τ can be derived by considering the following set of equations

166 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

xt();;;βββ=+ axt ()3 yt () a.s.tt≥ 0 (63) xt()()()+=+τβ;;; axt τβ3 ++ yt τβ

Now using same arguments as for the case of the derivation of the joint characteristic function

φυxy (),;tvt , we have the equation

∂∂∂φφφ()ts,,,33() tt() tt xxyy++aaυυ xxyy xxyy = ∂∂∂t 12υυ33 st= 12 ∂∂φφxxyy()tt,, xxyy () tt =+υυ12,,,,, υυ 1212vv ∈∈ t I ∂∂vv12

where for simplicity, we write the joint characteristic function φυυxxyy ()12,;tt ,++ τ ; vsvs 12 , ; , τ as φxxyy (ts, ). Moreover the additional condition holds

φττφτxxyy()()0,t ;0, t++=+= ; vsvs12 , ; , yy vsvs 12 , ; ,

=⋅−+⋅−−∈Y y ()vsvs12δδτ()ii ( ),, vv 12 and the initial condition

φυυτxxyy()()()12,0; , ;0,ss ;0,+= τφυυτ xx 12 ,0; , =Y 012 υυ , , υυ 12 , ∈

Similarly we can derive equations for joint characteristic functions of higher derivatives.

– Derivation of a P.D.E. for the joint characteristic function φxy (υ,;ttv ,) for a general system.

For the general case we have the following

Theorem 3.17. : Let the dynamical system Σ1 described by (18)-(19) of Section 5.2.3. Then the FDE

11d KL ⎡⎤∇=∇∇∇∇()uv,,,,kl F (t x y ) l k ()uv ⎣⎦⎢⎥ut()YYxy∑∑{} y x00 kl+ ut() vt () xy idtkl==00 i

N Y xy ()()υυ⋅=δ00,0 Y , υ ∈

∞,N ∞,M for every u ∈CIc (), v ∈CIc () and tI∈ , which describes the system, can be reduced to the partial integrodifferential equation

∂φ υν NL xy ()ts,;, υn ll +∇{}xxyFtn (),0,0{} ∇υφ () t ,υν ; s , = ∂ti∑∑ l+1 st= nl==10 (64) NKL υ =∇∇∇∇n kl klφ υν ∑∑∑ kl+−1 {}yxFtn (),0,0{}ν υ xy() t , ; s , nkl===110i with the additional condition, 5.3. REDUCTION OF THE HOPF-TYPE FUNCTIONAL DIFFERENTIAL EQUATIONS FOR SPECIAL CASES 167

φφxy()()()ts,;,0 ννν==⋅−∈ y s ,Y y () δi s , ν (65) and the initial condition

φφxy()()()0,υυυυ ;s ,0 == x 0,Y 0 , ∈ (66)

NM where φxy ()iiii,;, :II×××→ is the joint characteristic function of the system response x()t ;β and the excitation y()t ;β .

Proof : We consider the component wise form of the FDE, used for the proof of Theorem 2.1. (1,2,,nN= … )

KL kl+ kl+ 11d δδYYxy ()uv,,∂ Ft(),0,0 xy () uv =⋅n ∑∑kl+ ∑ (67) idtδδδ un () tkl==00 i()ij, ∈×IINM ∂∂xyij u i() t v j () t ij=∧lk = We separate the r.h.s. term as follows,

LKLl l kl+ kl+ 11∂∂Ft(),0,0δδYYxy ()uv,, Ft() ,0,0 xy () uv n ⋅+n ⋅ ∑∑lkl ∑∑+ ∑ lkl===010itittiij∈∈×IIIN ∂∂∂xuiiδδδ() (), NM xyuv ijij() () i =l ij=∧lk =

Setting to the system of functional differential equations u =⋅−()()υυ12,,…i υN δ t and v =⋅−()()νν12,,…i νM δ s we have using same arguments as for the proof of Liouville’s equation,

1 d δY xy ()υν⋅−δδ()iits, ⋅− () lim = st→ idtδ un () t NM 1 ⎪⎪⎧⎫ =+limixtυυνnn () exp⎨⎬ i xtκκ () i ys κ () κc xy ( ddx , y ) st→ ∫ ⎪⎪∑∑ iυn ∞∞,,NM ⎪⎪κκ==11 CCcc× ⎩⎭

For the first r.h.s term we have,

L l l 1 ∂ Ft(),0,0 δ Y xy ()υν⋅−δδ()iits, ⋅− () lim n ⋅= st→ ∑∑l l=0 iti∈I N ∂xuiiδ () i =l

L l l 1 ∂ Ftn (),0,0 ∂ φxy ()υν,;tt , =⋅∑∑l l=0 i i∈I N ∂∂xiiυ i =l

Similarly for the second r.h.s. term we have,

KL kl+ kl+ 1 ∂ Ft(),0,0 δ Y xy ()υν⋅−δδ()iits, ⋅− ( ) lim n ⋅= st→ ∑∑kl+ ∑ kl==10itt()ij, ∈×IINM ∂∂xyijδδ u i() v j () ij=∧lk = KL1 ∂kl+ Ft(),0,0 ∂kl+ φ ()υν,;tt , =⋅n xy ∑∑kl+ ∑ kl==10i ()ij, ∈×IINM ∂∂xyij ∂υν i ∂ j ij=∧lk =

168 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

Where we have used the notation,

ii12 iN ∂=∂⋅∂xi x12xx… ⋅∂N

i ii12 iN x =⋅⋅⋅x12xx… N

Multiplying all terms with iυn and summing up for all nN= 1,2,… , , we will have for the l.h.s. term,

NNM ⎪⎪⎧⎫ limixtυυνnn () exp⎨⎬ i xtκκ ()+= i ys κ ( ) κc xy ( ddx , y ) st→ ∑∑∑∫ ⎪⎪ n===111∞∞,,NM ⎪⎪κκ CCcc× ⎩⎭ ∂∂φφ()ts,;,υν() ts ,;, υν ==lim xy xy st→ ∂∂tt s=t

Making the summation for the r.h.s. terms we have the partial differential equation,

∂∂φφ()ts,;,υνNLυ ∂l Ft(),0,0 l () υν ,;, tt xy =⋅+n n xy ∑∑l−1 ∑ ∂∂∂tiN x υ st= nl==10 i∈I ii i =l NKL υ ∂kl+ Ft(),0,0 ∂kl+ φ ()υν,;tt , +⋅n n xy ∑∑∑kl+−1 ∑ nkl===110i ()ij, ∈×IINM ∂∂xyij ∂υν i ∂ j ij=∧lk =

It is easy to verify conditions (65) and (66). ■

Remark 3.18 : The separation of the r.h.s. that we have made allows us to give the physical meaning for the above equation. Thus, the l.h.s. of equation (64), which is equivalent with the l.h.s. of Liouville equation express the conservation of ‘probability’ for the state space variables x()t ;β . The r.h.s term, which is due to the excitation term, express the transfer-interaction of ‘probability’ from the stochastic process y()t ;β to the state space variables x(t ;β) . Of course the stochastic characteristics of the excitation y()t ;β are introduced with condition (65). ■

5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 169

5.4. Kernel representations for the Characteristic Functional

We will now study methods for the efficient numerical solution of the Functional Differential Equations described at Section 5.4.2 of the present chapter. These methods, in contrast to the previous section (based on the reduction of the FDE’s to PDE’s) will be based on analytical calculations of functional integrals (EGOROV, A.D. & SOBOLEVSKY, P.I. & YANOVICH, L.A., Functional Integrals: Approximate Evaluation and Applications) presented below. More specifically, using the linearity property of the Functional Differential Equations, we will apply a convex superposition of characteristic functionals (kernel characteristic functionals). Then two different approaches will be formulated. The first one will superpose kernel characteristic functionals with given parameters, efficiently chosen, such that the corresponding measures be local in space and time, i.e. in N . We will explain this property extensively in what follows (Section 5.4.3). Thus we will get a Gallerkin type approximation at functional level. The second approach will work in two stages. In the first stage we will approximate the initial condition distribution from a superposition of kernel characteristic functionals. In the second stage we will let the characteristic functionals to evolve in time with constants amplitudes (those that we computed). After an efficiently chosen time interval we will recompute the amplitudes, by approximating the computed probabilistic stage as it was an initial condition using suitable criteria.

5.4.1. Exact Formula for Integrals of Special Functionals

a) Integration with respect to Independent Increment Processes

We will present the above theorems for the more general case of X be a linear topological space (separable, locally convex linear space). X ′ will denote the dual space of linear continuous functionals on X . In what follows, we shall consider without any loss of generality the mean value of a Gaussian measure c to be equal to zero. In may cases, we may without any loss of generality consider the correlation operator of c ,Cxx( , ) to be nondegenerate, i.e., Cxx(),0= if and only if x = 0 (we have substitute the notation Cx, x of Chapter II, for the current suction, with Cxx( , ) to avoid confusions with the dual product). Otherwise, we could find the subspace X 0 ⊂ X where the measure is concentrated, and the correlation operator will no longer be degenerate there. The following construction is described and substained in papers (see EGOROV, A.D. & SOBOLEVSKY, P.I. & YANOVICH, L.A., Functional Integrals: Approximate Evaluation and Applications p.16) in more details. The Hilbert space which is the closure of the set of functionals of the form ξ,i , ξ ∈ X ′ in space _c2 ()X , will be denoted by H and [ ∈ X denotes the Hilbert subspace dual to H whose closure in X is the support of measure c . For almost all x ∈ X , a functional (ax, ) ()axX∈∈[ , is defined. It is specified by the series

∞ ()ax,,,= φ x() ae (1) ∑ kk[ k=1 170 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

where {}φk , {}ek , k =1, 2,…, are orthonormal bases in H and [ , respectively, with

φk ∈ X ′ for all k and φδkj,e = kj. We assume that the space [ is seperable for all spaces X and measures c considered. In particular, this is true for separable Frechet spaces and for other spaces considered here. Functionals of the form (1) are called measurable linear functionals. Note that ∞ ()ex,,= φ x , ()ah,,,,==φ x()() ae ah kk ∑ kk [ k=1 for ah, ∈[ . It is an important property of the considered spaces with Gaussian measure that they admit the following expansion

∞∞ x ==∑∑()exekk,, ()φ kk xe (2) kk==11 which converges under topology of space X for almost all x ∈ X . Proofs of the convergence of expansion (2) may be found in papers. The definition of H implies that ()ξη,,,,== ξxx η() dxC ( ξη ). Let T define an isomorphism of space into H H ∫ c [ which assigns a basis {}φk to the basis {}ek . Then

∞∞ −−11 ξξφφξφξ,,,x ==()kkxTexTx() ,,, k k =() (3) ∑∑H [ kk==11

∞∞ −−11 −− 11 CTeTeTT()ξη,,,== ( ξφkk )() ηφ()()() ξ , k η , k = ξ , η ∑∑HH [[ kk==11

For the proofs of the theorems presented below we will need the following transformation formulae for the integrals with respect to Gaussian measure under translation and under general linear transformation may be obtained. Thus, suppose F ()x is a functional to be integrated and a∈[ . Then the transformation formula for an integral with respect to Gaussian measure under translation is as follows

⎪⎪⎧⎫1 2 F() xcc ( dx )=−exp⎨⎬ a F() x + a exp{} − () a , x() dx (4) ∫∫⎪⎪[ XX⎩⎭⎪⎪2

In order to state the next result, we need some definitions. A A in Hilbert ∞ 2 space is called a Hilbert-Schmidt operator, if ∑λk <∞, where {}λk ()k =1, 2,… is the k=1 ∞ sequence of its eigenvalues of A . On the other hand, if the series ∑ λk <∞ converges then k=1 this operator is called the operator of . Clearly, any trace class operator is Hilbert- ∞ Schmidt one. The product DA ()λλλ=−∏ (1 k ), where {}λk ()k =1, 2,… is the sequence of its k=1 eigenvalues of A , is called the Fredholm determinant at point λ for operator A . The Fredholm determinant of a trace class operator is finite for any λ . It is an entire function of −1 complex variable λ with zeroes at points λλ= k ()k =1, 2,… . If A is a Hilbert-Schmidt 5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 171

∞ operator then the product ∏()1−λλk may diverge. For such operators, the determinant of the k=1 ∞ λλk form δλAk()=−∏ (1 λλ )e is introduced and is called the Carleman determinant. k=1 Let us proceed now to the evaluation of integrals with respect to Gaussian measure for particular functionals. First we shall evaluate integrals of functionals which are functions of measurable linear functionals

Theorem 4.1 : Let aa1,,… n be linearly independent elements from [ . Then the following equality holds

−−n /2 1/2 ⎪⎪⎧⎫1 −1 Fax()()()1,,,… axn ,c ()( dx=− 2π ) [ det A ] exp⎨⎬() Auu , F() u d u (5) ∫∫⎪⎪2 X n ⎩⎭⎪⎪

⎡⎤ if any of these integrals exists, where A is a matrix with elements Aaaij= () i, j ⎣⎦⎢⎥ [ ()ij,1,2,,= … n.

Proof : We shall make use of formula (4) for the transformation of integrals under translation, n wherein we set aa=−λυ∑ kk, with λ and υk be real numbers, and Fx()=1. Then we have k=1

nn⎪⎪⎧⎫222 ⎪⎪⎧⎫ ⎪⎪ ⎪λλ ⎪⎪⎪⎪⎪⎧⎫ exp⎨⎬λυ()()ax ,c dx== exp ⎨ υ a ⎬⎨⎬ exp() A υυ , ∫ ⎪⎪∑∑kk ⎪22 kk ⎪⎪⎪ X ⎩⎭⎪⎪kk==11⎩⎭⎪⎪[ ⎩⎭⎪⎪

This equality is also valid for λ =−i . This together with the following formula

⎪⎪⎧⎫11n/2 1/2 ⎪⎪⎧⎫ exp⎨⎬⎪⎪−−=()Aid−1uu ,() u ,υυυ u ()[] 2π det A exp ⎨⎬⎪⎪ −() A , ∫ ⎪⎪22 ⎪⎪ n ⎩⎭⎪⎪ ⎩⎭⎪⎪ implies that

⎧⎫n ⎪⎪−−n/2 1/2 ⎪⎪⎧⎫1 −1 −i()u,υ exp⎨⎬−=iaxdxυπkk()()()[] ,c 2 det A exp ⎨⎬ −() Auu , ed u ∫∫⎪⎪∑ ⎪⎪2 X ⎩⎭⎪⎪k=1 n ⎩⎭⎪⎪ i.e. formula (5) is valid for Fe()u = −i(u,υ) . Let now the function be specified by its Fourier transform, i.e. Ffed()u = ∫ ()υυ−i(u,υ) . Let us multiply the previous relation by f (υ) and n n integrate it over space with respect to the variables υυ12,,,… υn . After changing the order of integration we obtain (5). The validity of the formula for arbitrary functions F (u) is verified by passage to the limit. ■

Remark 4.2 : For the case of integration with respect to a Gaussian measure with non zero mean value m we have Fax,,,…… ax , dx=+ Fax , am , ,, ax , + am , dx = ∫∫()()()111nnncc()() ( )( ) ( )( ) () XX

−−n /2 1/2 ⎪⎪⎧⎫1 −1 =−++()2detexp,π [AAFuamuamd ] ⎨⎬()uu()11() ,,,,… nn () u ∫ ⎪⎪2 n ⎩⎭⎪⎪ 172 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

where c , is the Gaussian measure c with zero mean value. Making the suitable translation transformation to the last integral we have

Fax,,,… ax , dx= ∫ ()()()1 n c () X −−n /2 1/2 ⎪⎪⎧⎫1 =−−−()2detexpπ [AAmmFd ] ⎨⎬⎪⎪()−1 ⎡⎤⎡⎤ua() ,,, ua () () uu ■ (6) ∫ ⎪⎪2 ⎣⎦⎣⎦ n ⎩⎭⎪⎪

⎡⎤ Remark 4.3 : For case of X be a separable real Hilbert space, matrix A with Aaaij= () i, j ⎣⎦⎢⎥ [ will take the simple form ⎡⎤ACaa= , . ■ ⎣⎦⎢⎥ij() i j

Let us mention two special cases of the obtained formula. For n =1 we have

⎪⎪⎧⎫2 −−1/2 1 ⎪⎪u Fax()(),2expc ()( dx=−π ) a⎨⎬ Fudu() ∫∫[ ⎪⎪2 a 2 X ⎩⎭⎪⎪[

n Let aa12,,,… an be orthogonal elements of space . Then

−n /2 ⎪⎪⎧⎫1 Fax()()(),,,… ax ,c ()( dx=− 2π ) exp⎨⎬()()uu , F u d u ∫∫1 n ⎪⎪2 X n ⎩⎭⎪⎪

Consider now evaluation of integrals of functionals which are functions of quadratic functionals on X (for example the Tatarskii representation)

∞ Axx(),,,= ∑ akj() e j x() e k x (7) kj,1= where aAeekj=() k, j ; A is a self-adjoint trace class operator on [ ; {}ekk ,1,2,,= … is an orthonormal basis in [ . The double series in (7) converges for almost all x ∈ X . The functional Axx( , ) does not depend on the choice of basis {}ek and may be written as follows

∞ 2 Axx(),,= ∑ λkk ( e x ) kj,1=

where {}λk and {}ek are the eigenvalues and the orthonormal eigenvectors of operator A , respectively. The following integrals are easy to evaluate with the help of (5)

∫ Axx()(),tracec dx= [] A X 222⎡ ⎤ ∫ A ()()xx,c dx=+ trace [] A 2trace ⎣⎢ A⎦⎥ X

Let us prove the equality

⎡⎤λ −1/2 exp⎢⎥Axx()() , c dx= ⎡⎤ D ()λ ∫ ⎢⎥ ⎣⎦A X ⎣⎦2

5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 173

−1 where DA ()λ is the Fredholm determinant of operator A at point λ ; Reλλ< 1 (λ1 is the largest eigenvalue) and

−−1/2 1/2 ⎪⎪⎧⎫i ⎡⎤λλ=−⎪⎪ λ ⎣⎦DAA() D () exp⎨⎬ argD A() ⎩⎭⎪⎪2 In fact, ∞ ⎡⎤λλ⎡ 2 ⎤ exp⎢⎥Axx()() ,cc dx== exp⎢⎥λ () e , x () dx ∫∫⎢⎥22⎢⎥∑ kk XX⎣⎦ ⎣ kj,1= ⎦ ∞∞⎡⎤ −−1/2⎪⎪⎧⎫1 1/2 −1/2 =−=−=⎢⎥()2exp1πλλλλλ⎨⎬⎪⎪()udu2 () 1 ⎡ D ()⎤ ∏∏⎢⎥∫ ⎪⎪2 kkA⎣ ⎦ kk==11⎣⎦⎢⎥ ⎩⎭⎪⎪

We may similarly evaluate an integral of the more general form

2 −1/2 ⎪⎪⎧⎫b exp⎡⎤⎡⎤λλλAxx()()() ,+= bgx ,c dx D () 2 exp⎨⎬() B() 2 gg , (8) ∫ ⎣⎦⎣⎦AA⎪⎪2 [ X ⎩⎭⎪⎪

−1 −1 where g ∈[ , BA ()(212λλ=− A ), Reλλ<( 2 1) (λ1 is the largest eigenvalue); b is a numeric parameter; I is the identity operator. Now formula (8) will be exploited in the proof of a more general result.

Theorem 4.4 : Let Axx1 (),,,… Axxn () , be quadratic functionals of form (7) and let

gg12,,,… gm be linearly independent elements of space [ . Then the following equality holds

FAxx,,,…… A xx ,; gx ,,, g , x dx= puuu ;υυυ F ; dd (9) ∫∫()11()nm ()() ( )c () ()() X nm+ (if any of these integrals exists), where

−−nm/2 −1/2 ⎪⎪⎧⎫1 −1 p()()uu;υξξξυυξ=−− 2π exp{}iDiS () ,⎡⎤() 2 det⎡⎤ () exp⎨⎬⎪⎪ S() , d ∫ ⎣⎦⎢⎥A()ξ ⎣⎦ ⎪⎪2() n ⎩⎭⎪⎪ −1 −1/2 S ξ is a matrix with elements ⎡SIiAggξξ⎤ =+⎡⎤2,; ⎡⎤D λ is the () ⎢⎥kj()()⎣⎦ () k j ⎢⎥A()ξ () ⎣⎦ [ ⎣⎦ n Fredholm determinant of the operator AA()ξ = ∑ξkk at point λ ; u =()uu1,,… n and k=1

υ =()υυ1,,… m .

Proof : The proof of this theorem is similar to that of Theorem 4.1, i.e., first we prove the validity of formula (9) with the help of formula (8) for the functions F ()u;υ which are specified by their Fourier transforms; and then we prove the formula for arbitrary functions by passing to the limit. ■

Remark 4.5 : As we notice at Remark 4.2 for the case of integration with respect to a Gaussian measure with non zero mean value m we have

FAxx,,,…… A xx ,; gx ,,, g , x dx= ∫ ()11()nm ()() ( )c () X =+++++FAxmxm,,,,;,,,,…… Axmxmgxm gxm + dx = ∫ ()11()()()()nmc () X 174 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

∫ p()uuu;;υυυFdd () nmn++ where c , is the Gaussian measure c with zero mean value and we have apply Theorem 4.4 for FFAxxAmxAmmAxx=++( 111(),2, ( ) ( ,,,, )… n () +

++2,Amxnn() Ammgx ( ,;,,,, )(1 + m )(… g m x + m )) ■

Remark 4.6: For case of X be a separable real Hilbert space, matrix S will take the simple −1 form ⎡⎤SCIiAggξξ=+⎡⎤2, where C is the covariance operator of the Gaussian ⎣⎦⎢⎥kj()()⎣⎦ () k j measure. ■

Consider special cases of formula (9). If F ()u;υ does not depend on u , i.e., FF()u;υυ= (), then we obtain formula (5). Suppose that FF()uu;υ = (), i.e., it does not depend on υ , then

FAxx,,,… A xx , dx= ρ uuu F d ∫∫()1 ()n ()c ( ) () () X n where −1/2 ρπ()uu=− (2exp, )−n iDid (ξξ )⎡ () 2⎤ ∫ {}⎣⎦⎢⎥A()ξ n n

DiA()ξ ()−2 is the Fredholm determinant of the operator AA()ξ = ∑ξkk at point λ =−2i . k=1 Consider still another example ⎧⎫2 ⎡⎤λυ−1/2 ⎪⎪ exp⎢⎥Axx()() , F() gx ,c () dx=−⎡⎤ 2πλλ D ()() b exp⎨⎬ F() υυ d ∫∫⎢⎥22⎣⎦A ⎪⎪b λ X ⎣⎦ ⎩⎭⎪⎪() where bIAgg()λλ=−() [ ]−1 , . [

b) Integration with respect to Independent Increment Processes

For the completion of the Section we also report some useful formulas for the calculation of integrals with respect to measures which correspond to independent increment processes. Measures of this kind are featured by integrals of the form

Fx()()i…=Δ f() xt()1 ,, Δ xt ()n

where Δ=x()txtxtkkk () − (−1 ). Thus for the measure which corresponds to Poisson process with the characteristic functional

⎪⎪⎧⎫ ⎪⎪⎡⎤iuxαρt (), Y ()x =−exp⎨⎬λ edu 1 ⎪⎪∫ ⎣⎦⎢⎥ ⎩⎭⎪⎪T where ⎪⎧1, iftuT∈[] , ρ ()u = ⎨⎪ , t ⎪ ⎩⎪0 , otherwise 5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 175

the following formula holds

k ∞ n ()λΔt i fxt()ΔΔ(),,…… xt ()c () dx = fkk ( ,, ) i e−Δλ ti (10) ∫ 11nn∑ ∏ k ! X kk1 ,,… n = 0 i=1 i and for the measures generated by the infinite-dimensional Gamma distribution with the characteristic functional

⎧⎫ ⎪⎪ Y ()x =−exp⎨⎬⎡⎤ ln 1iuxduσρ () , ⎪⎪∫ ⎣⎦⎢⎥()t ⎩⎭⎪⎪T where ⎪⎧1, iftuT∈[] , ρ ()u = ⎨⎪ t ⎪ ⎩⎪0 , otherwise the following formula is valid

Δ−t 1 u n ⎡ u k − k ⎤ ⎢ k σ ⎥ f ()ΔΔxt()11,,…… xt ()nnc () dx = f ( u ,, u ) e du (11) ∫∫∏⎢ Δtk ⎥ n k=1 ⎢ΓΔ()tk σ ⎥ X + ⎣ ⎦ ∞ where Γ=y ueduyu−−1 is Euler’s Gamma function, n is the product of the 4th positive () ∫ + −∞ half axes. We shall also give an example for the case of the characteristic functional ⎪⎪⎧⎫11⎡⎤2 p ⎪⎪ Y ()x =−exp⎨⎬⎢⎥x () t dt du (12) ⎪⎪∫∫⎢⎥ ⎩⎭⎪⎪0 ⎣⎦⎢⎥u which is the special case of the characteristic functional of form (22) of Chapter 3/Section 3.9 2 p for T =1, σ ()2!p , ν ()du= du and ρt (u) as we defined it before. Note that (10) may also be written in the form

⎧⎫11 ⎪⎪ Y ()x =−exp⎨⎬………… mint , , t x() t x t dt dt (13) ⎪⎪∫∫ ()12ppp 1() 212 ⎩⎭⎪⎪0 u

The following formula holds

n f ΔΔxt,,… xt dx = fuu S Δ− t , u u d (14) ∫∫()()1 ()nkkkc () ()∏ (−1 ) X n k=1 where

1 Su()ττυυυ,exp=−+{}2 p iud π ∫ 2 is the fundamental solution of the parabolic equation

2 p ∂∂SSp+1 =−()1 ∂∂τ u2 p

Using (12) we obtain the following formula

176 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

⎛⎞11 ⎜ ⎟ faxd⎜ 1 ()()τττ,,… aν ()() τττ xd⎟c ( dx )= ∫∫⎜ ∫ ⎟ X ⎝⎠00 ⎧⎫2 p ⎪⎪⎪⎪⎧⎫11⎛⎞n ⎪⎪−n ⎪⎪⎜ ⎟ ⎪⎪ =−−f ()(uu⎨⎨2exp,πυτ )iasdsddd (υυ )⎜ () ⎟ ⎬⎬ u ∫∫⎪⎪ ∫∫⎜⎟⎜∑ kk ⎟ ⎪⎪ nn⎪⎪0 ⎝⎠k=1 τ ⎪⎪ ⎩⎭⎪⎪⎩⎭⎪⎪

n where ()u,υ = ∑ukkυ , which is valid under the condition of measurability of f ()uu1,,… n k=1 and the condition

⎪⎪⎧⎫p ⎪⎪21p− f ()()uu11,,……nn exp⎨⎬−≤ε uu , Huu() ,, ⎩⎭⎪⎪ for any ε > 0, where

Hu,,…… u du du<∞. ∫ ()11nn n For more explicit formulas for the calculation of infinite dimensional integrals we refer to EGOROV, A.D. & SOBOLEVSKY, P.I. & YANOVICH, L.A., Functional Integrals: Approximate Evaluation and Applications.

5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 177

5.4.2. Integrals of Variations and of Derivatives of Functionals.

Let functional Fx() be differentiable along direction a ∈[ at any point x ∈ X and let

d sup Fx()+λ a λε< dλ be summuble for some ε > 0. Then the following relation holds

∫∫δFxa()[]cc ( dx )= Fx ()( ax, ) ( dx ) XX

The latter relation follows from the identity

⎧⎫2 ⎪⎪λ 2 Fx()()+=−λλ acc dxexp⎨⎬ a Fx() exp{} ( ax , ) () dx ∫∫⎪⎪2 [ XX⎩⎭⎪⎪

(which is a modification of formula (4)), if we differentiate it with respect to λ and set λ = 0 . Moreover the above condition enables differentiation under the integral sign.

The functionals Hermite polynomials may be defined by the equality

n ()n ∂ Hxaaa[];,,12…… ,nn= Gxaaa[] ;,, 12 , ∂∂λλ12… ∂ λn λλ1==… n =0 where

⎡⎤nn ⎢⎥1 Gxa[];,,12 a… , anijijii=− expλλ() a , a + λ () a , x ⎢⎥∑∑[ ⎣⎦2 ij,1== i 1

ai ∈[ ; the functional Gxa[];,,12 a… , an is defined for almost all x ∈ X . The following recurrence relation holds for functional Hermite polynomials

(nn) ( −1) Hxaaa[]();,,12…… ,nn=− axHxaaa ,[] ;,, 12 , n− 1 n−1 − ()aa,;,,,, H()n−2 [] xa…… aˆ a ∑ in[ 11 i n− i=1

(here the “hat” over ai means that ai must be omitted); it is verified by an immediate calculation with the help of the above-mentioned definition of functional Hermite polynomials. The explicit form of them is as follows.

s n ⎣⎦⎢⎥n/2 ()n ()−1 Hxaaa[]();,,12… ,ni=+ ax , ∑ × ∏ ()(2!!sn− 2! s ) i=1 s=1 n sn × aa,, ax ∑ ()()jj21pp− 2 j 2 q ∏∏[ jj1≠≠=… n 1 pqs==+121 where ()2!!24ss=⋅⋅… ⋅ 2, ⎣⎦⎢⎥i is the floor function. The first three Hermite polynomials are Hxaax(1) []();,= ,

Hxaa(2) []()()();,=− axaxaa , , , , 12 1 2 12[ 178 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

H(3) [ xaaa;,, ]()()()()()=−− axaxax , , , aa , ax , 123 1 2 3 13[ 2 , −−()()()()aa,,,, ax aa ax 12[[ 3 23 1

Note also the special case of aa12====… aan

j ⎣⎦⎢⎥n/2 ()−1!n 22j nj− HxaHxaaa()nn[];;,,,==()[]… aax() , ∑ [ j=1 ()(2!!2!jn− j )

The following formula holds

δn Fxaa,,,…… a dx= FxH(n) xaa ;,,, a dx ∫∫()[12nn ]cc ( ) () []() 12 XX provided that these integrals exist and

∂ ⎛⎞n sup Fx⎜ + λ a⎟ ⎜ ∑ ii⎟ λε< ∂∂λλ1 … n ⎝⎠i=1 is summable for some ε > 0. The formula follows immediately from the definition of the n-th order variation, the definition of Hermite polynomials given at the beginning of the section, and relation (4) or its modification also given at the beginning of the section.

5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 179

5.4.3. Kernel Characteristic Functionals

We shall now use the theorems for analytical calculations of infinite dimensional integrals presented above. As we mention at the introduction of the section we will use a convex superposition of kernel characteristic functionals, i.e. functionals that correspond to probability measure which are local into the functional space (kernel probability measures). These analytical calculations, will allows us to get a system of differential equations for the parameter functions of the kernel characteristic functionals. First we will give a more precise definition of kernel characteristic functionals and some simple examples. Then we will illustrate the proposed method, by applying to a first order equation with cubic nonlinearity, which has been studied at Section 5.3.4 of the present chapter. In following section, we will also present the theorems concerning the approximation of an induced probability measure of finite dimension (density) from a kernel type representation. Thus, let us introduce a precise definition of kernel characteristic functionals

Definition 4.7 [()ε,h − Kernel Characteristic Functionals] : Let [ be a linear topological space(1) which will be mainly thought of as a space of functions defined on the time interval I . Let also U[() denote the σ −field generated by the open subsets of [ (the Borel

σ − field of [ ), m ∈[ , h ∈ + and h mh, be a neighbourhood of m with concentration parameter (e.g. radius, in the case of metric spaces) h . [For example, h mh, can have the form h mh, = {}uumh∈−≤[ : if an appropriate norm is available.]

Given ε > 0, we shall say that a characteristic functional Y (umh;,) is an (ε,h)−kernel characteristic functional(2) iff the associated probability measure c on U[() is concentrated around the element m ∈[ in the sense that

c ()h mh, ≥−1 ε . ■

The existence of (ε,h)−kernel ch. functionals, for any given ε > 0 and h > 0 (relatively small), will be established in Section 5.4.3a by means of construction of a specific example.

Remark 4.8 : To understand the above definition we study the space of real valued continuous functions with the sup norm defined on the time interval I =[]0,T . Then an arbitrary set of the form = uutmth∈−≤:sup () () will be a set of functions that take values within h mh, {}[ tI∈ the shaded region

(1) Reflexive sometimes Banach or Hilbert (2) The terms ε−kernel ch. functional or –simpler– kernel ch. functional will also be used if no confusion is likely to occur. 180 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

x(t)

= uumh∈−≤:sup h h mh, {}[ mt( ) tI∈

0 t T As h → 0 the above region will be reduced to the element m . The form of the characteristic functional/probability measure is evolved from the parameter h (bandwidth) which adjust the variance of the probability mass about the element m . Usually the parameter h is taken us the ⎡⎤1/2 maximum value of the variance, ⎣⎦Ctt(), and m as the mean value element of the probability measure. ■

The notion of ()ε,h − kernel ch. functionals will now be exploited to establish a fundamental result concerning the approximation of any probability measure (at the level of projection into finite-dimensional subspaces). More precisely, we shall prove that any finite-dimensional projection of a given probability measure c on U[(), can be approximated by the corresponding projection of an appropriate convex superposition of kernel ch. functionals.

Let c[ be a given probability measure defined into the measurable space ()[U[, () and let ΠL be a projection of [ into the finite-dimensional subspace L . We then have the projection (See Definition 1.6/Chapter 3) B ==∈∈∈ΠΠ−1 Bx: xBBL, ccΠL () [[()LL[ ] c[({ [ ] }) U() which is a probability measure defined into the finite dimensional subspace L . Let also the set of all finite linear combinations (convex) of the form

n n cc()ii= ∑ pkkk();,mh, with pn ≥ 0 , nN=1,… , and ∑ pn = 1 (15) k=1 n where every c ()i;,mhkk ()kn=1,… , is a kernel probability measure. Similarly we have the projection of the kernel probability measure nnB ==∈∈∈ΠΠ−1 Bx n: xBBL, ccΠL () ()LL[ ] c[({ [ ] }) U() We then have the following

Theorem 4.9 : For every projection ΠL from [ into the finite dimensional subspace L , the set of measures n B is dense into the set of all continuous with the same support. c ΠL ( ) c ΠL Proof : See ATHANASSOULIS, G.A. & GAVRILIADIS, P.N., 2002, The Truncated Hausdorff Problem solved by using Kernel Density Functions. Probabilistic Engineering Mechanics, 17, 273-291. ■

5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 181

Remark 4.10 : In other words, for every measure c[ ()i , which has continuous projection into the finite dimensional subspace L and every ε > 0, given a kernel measure k , c ΠL c there is a finite set of mean value elements mk ∈ supp[]c[ , and real numbers hk and pk ()kn=1,… , satisfying conditions (15) such that

B −<∈n BBLε , ■ (16) ccΠΠLL() () U( )

Remark 4.11 : By assuming the necessary continuity conditions we can have the corresponding results for the characteristic functional. ■

Remark 4.12 : We must note that the kernel property of a probability measure pass along to all of its projections on finite dimensional subspaces in the sense that for every projection ΠL the probability measure has the kernel property, i.e., c ΠL

hh==∈∈≥−ΠΠ−1 ⎡⎤xB: ⎡⎤ h1 ε . ■ ccΠL ()mh,,() L⎣⎦ mh c[({ L ⎣⎦ mh ,})

182 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

5.4.3a Gaussian Kernel Characteristic Functionals

The most immediate application of the above definition is for the case of Gaussian characteristic functionals. Functionals of this type have the advantage of an explicit form for both the characteristic functional and the probability density functionals. Suppose we are working in a suitable space of functions defined on the time interval I =[]0,T . Then the following theorem due to D.G. Konstant and V.I. Piterbarg (Extreme values of the cyclostationary gaussian random process) gives a set of sufficient conditions for a Gaussian probability measure to be of kernel type with the sup norm.

Theorem 4.13 : Let X ()t ;β , tI∈⊆ IR, be a Gaussian random process which is differentiable in mean square, with (differentiable) mean-value function mI(i): → and set 2 dEXtmt2 =−>β ⎡⎤′′;0β . The covariance function ii×→ satisfies the t ()⎣⎦() () CII(),: strict inequality Cts(),,<≠σσ ()() t s t s, and the variance σ 2 ()tCtt= (, ) reaches its 22 maximum σσ()t00= at a unique point tI0 ∈ . Moreover, we assume that there exist two 2 constants, Α>0 , b > 0 , defining the local structure of σ ()tCtt= (, ) at the vicinity of tI0 ∈ such that bb σσ()tttotttt=−Α−+00() − 0, → 0 Then b ≥ 2 (because of the differentiability), and 1. If b > 2 , then c uutmth∈−>[ :max () () = ()tI∈

2 2 ⎛⎞1 dh− ⎛⎞ ⎜ ⎟ 0 b ⎜ ⎟⎡⎤ =Γ⎜ ⎟ 11hohexp⎜ −2 ⎟ 1 +() 1 , →∞ , ⎝⎠⎜ ⎟ 1+ ⎜ ⎟⎣⎦ b bb ⎝⎠2σ0 πσbΑ 0 2. If b = 2, then c uutmth∈−>[ :max () () = ()tI∈

⎛⎞⎛⎞d 2 h2 =+σ ⎜⎜1exp11,,0 ⎟⎟hoh−1 −+⎡⎤() →∞ 0 ⎜⎜⎟⎟2 ⎣⎦ ⎝⎠⎝⎠⎜⎜22Ασσ00 2 where dEXt22=−=>β ⎡⎤′′();0β mt () d . 000()⎣⎦t0 Proof : For the proof we refer to the paper of KONSTANT, D.G. & PITERBARG, V.I., 1993, Extreme values of the cyclostationary gaussian random process. Journal of Applied Probability. ■

Remark 4.14 : Note that the condition of differentiability is crucial. For example, if consider any stochastic process with no memory at all (see Chapter 2, Section 2.2.1.a) then the probability of the maximum will be 0 or 1 (0 or 1 Kolmogorov Law). To illustrate the above consider a white noise process with constant intensity D . Then the maximum between n time instants will be 5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 183

⎛⎞n ⎜ 2 ⎟ hh ⎜ u ⎟ ⎛⎞1 ⎜ ∑ k ⎟ ⎜u∈>:max u t h⎟ = …… exp⎜ −k=1 ⎟ du du = c ⎜ [ () ⎟ ∫∫ n ⎜ 2 ⎟ 1 k ⎝⎠tt∈{}1 ,,… tn 2 n ⎟ −∞ −∞ 2π D ⎜ 2D ⎟ () ⎜ ⎟ ⎝⎠⎜ ⎟ ⎛⎞n ⎜ u2 ⎟ ∞ ⎜ ∑ ⎟ 1 ⎜ k=1 ⎟ m = 1 exp⎜−=→2 ⎟du ε 0 , as m →∞. ■ ∫ 2 n ⎜ ⎟ h 2π D ⎜ 2D ⎟ () ⎜ ⎟ ⎝⎠⎜ ⎟

Hence for the case of a Gaussian measure as described at Theorem 4.13 we will the description by the family of one-dimensional distributions, 2 ()tmx− , a 1 − ayxya=∈;, ≤= edt2,Cx x , x ∈ , a ∈ cc[x () G {}∫ [ −∞ 2,π Cx x Setting to the above x()ii=−δ (t ) we have the probability of values at a specific time t,

2 ()tmt− () a 1 − 2,Ctt() ∈ ∈ c[G {}y ∈≤=; yt() a∫ e dt , a , tI −∞ 2,πCtt() Thus, we see that ƒ the probability mass is concentrated to functions that have small distance from the mean value element mI: → . By ‘small distance’ we mean that sup()x()tmth−≤ () 2 . tI∈ Moreover we see that for h → 0 the probability is concentrated to the mean value element. ƒ For a convex superposition the probability mass is shared between the supports of the 11 Gaussian measures. More precisely let the convex superposition cccGG=+ G 2212 with the Gaussian measures , described by the parameters mm, and CC, ccGG12 12 12 respectively. This kind of description is very essential for the representation of nonlinear dynamical system’s response. x(t)

⎡ ⎤1/2 maxCtt1 () , tI∈ ⎣ ⎦

mt1 ( )

⎡ ⎤1/2 maxCtt2 () , tI∈ ⎣ ⎦ 0 T t

mt2 ()

■ 184 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

We must note that for the case described above, we have more precise information for the description of the probability mass. This is because we have the exact variation (through the ⎡⎤1/2 variance ⎣⎦Ctt(), ) from the mean element m for each time instant t . In other words we know that, given a covariance function with diagonal Ctt(), and the mean value element mt( ), the probability mass will be concentrated into the functional set

⎡ ⎤1/2 h = {}uutmtCtt∈−≤[ :,() () ⎣ ()⎦

The figure below illustrates graphically the above assertion for a given Ctt(), and mt( ). Ctt(),

t T

x(t)

t T

Another interesting subject connected with the above discussion is the characterization of the derivative of the stochastic process. In Chapter II we saw that having the characteristic function of a stochastic process gives us the opportunity to get statistical information for the derivatives of the stochastic process. Now we are going to discuss how we can get qualitative results for the probability mass of the derivative of the stochastic process, by having the behaviour of the correlation operator. We must say that the same arguments hold for a general stochastic process, so we can make the desired generalizations. Now, the correlation function of a mean-square continuous stochastic process can be expanded into a biorthogonal series of the form (Kahurent-Loeve expansion)

∞ Cts(), = ∑λii f ()() t f i s (17) i=1 ∞ where λ , ft( ) are the eigenvalues – eigenfunctions of the correlation operator {}ii i=1

Cx,, x= ∫∫ C()()() t s x t x s dtds II× Using the results of Chapter 3 we can have the distribution of the values for the derivative of the stochastic process 5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 185

Pa()=∈{} y; yta′ () ≤= () a Xt′′() c[0 cδ () tt− 0 Hence ⎪⎪⎧⎫ ⎪⎪′ 2 a ⎪⎪()tmt− () 1 ⎪⎪− ′ ⎪⎪2 c[G {}y ∈≤=;expyt() a⎨⎬∂ Cts(), dt , a ∈ , tT∈ ∫ 2 ⎪⎪ −∞ ∂ Cts(), ⎪⎪2 2π ⎪⎪∂∂ts ∂∂ts ⎪⎪st= st= ⎩⎭⎪⎪ ∂2Cts(), Thus we must investigate the behaviour of the function or using (17) the ∂∂ts s=t behaviour of the function

2 ∞ ∂ Cts(), 2 = λ ⎡⎤f ′()t ∂∂ts ∑ ii⎣⎦ st= i=1 Hence we can conclude that when the diagonal of the correlation function, Ctt(), change slowly with time, then the probability mass for the derivative of the stochastic process will concentrate very close to the derivative of the mean value element. Conversely, if the diagonal of the correlation function, Ctt(), change fast with time, even if decreases the probability mass will spread widely around the derivative of the mean value element. Ctt(), Ctt(),

t t T T

x′()t x′()t

T T t t

The above results, although easy to derive, gives us a very important physical intuition for the distribution of the probability mass into the infinite dimensional space of functions. Moreover, based on the analysis, presented for the simple case of Gaussian measures, one can produce valuable results for general functionals (for example Gamma characteristic functional), or for the responses of dynamical systems. Hence for responses having the form of convex superposition of Gaussian functionals, which will be used in the sequel, we can immediately have the picture of the probability measure. 186 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

5.4.4. Superposition of Kernel Characteristic Functionals for the numerical solution of FDEs

In the present section we will apply the representation described above for the numerical solution of functional differential equation. Theorem 4.9 in combination with the linearity of the Functional Differential Equations is the basic reason for the use of such representations to the numerical solutions of stochastic differential equations. In what follows we will see how we can get equations describing efficiently the evolution of the unknown parameters. Now let us describe in more detail the representation of the joint characteristic functional that we will use for the sequel. We assume that

a1. The random excitation is described by a stochastic process y(ii;:) I ×Ω → M defined onto the probability space (Ω, Uc(Ω), ) with the associated probability space ∞∞,,MM ()CIcc(),,Uc() C y , and with characteristic functional Y y . a2. The response is described by the stochastic process ()xy()()ii;, ii ; :I ×Ω → NM × defined onto the probability space (Ω, Uc(Ω), ) with the associated probability ∞∞,,NM ∞∞ ,, NM space ()CCcc××,,Uc() CC cc xy , and with characteristic functional Y xy . N a3. The initial state of the system is described by a random variable x0 ()i:Ω→ with MM the associated probability space (),,Uc()x with characteristic function Y 0 . 0 (Where we have assumed that for the initial state the random variables x()0;i :Ω→ N and y()0;i :Ω→ M are independent).

Consider the representation (15) of the joint probability measure using kernel probability 1 measures with fixed real numbers p ==,1,,kn… . k n n nkkkk1 ∞∞,,NM ccxy ()x,,;,,,y = ∑ xy ()x y mmhhx y x y , xy∈∈CCcc, (18) n k=1

kkkk with c xy ()x,;y mmhhx ,y ,x ,y being a kernel probability measure (Definition 4.7) and kN∞, kM∞, kNkM mx ∈ Cc , m y ∈ Cc , hx ∈ , hy ∈ for all kn= 1,… , . The characteristic functional corresponding to the above probability measure is

n nkkkk1 ∞∞,,NM Y xy ()uv,,;,,,= ∑Y xy () uv mx my hx hy , uv∈∈CCcc, (19) n k=1 with Y xy being the characteristic functional of the kernel probability measure c xy .

kkkk Remark 4.15 : Note that the parameters mmhhx ,,,y x y are not necessarily defining uniquely the kernel probability measures. In other words those parameters, express the kernel property of the probability measure c xy and they result from the specific parameters of the kernel kk representation. For example, in the case where c xy is a Gaussian kernel measure, mmx ,,y will kk be the mean value elements and hhx , y will be related with the correlation operator (more 5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 187

specifically with the variance vector). Hence, for the sequel we emphasize, that, when we refer kkkk to the determination of the unknown parameters mmhhx ,,,y x y we will mean the determination of the unknown parameters that define uniquely the kernel probability measure (for the Gaussian measure, those will be the mean value and the correlation operator) and hence they result the above quantities. ■

The characteristic functional (19) must satisfy the Functional Differential Equation governing the dynamical system, as well as, conditions a) and c) described above. Hence the characteristic functional must verify the conditions b1. Related with condition a1 for the characteristic functional of the excitation, we must have

n nn1 kk ∞,M YYxy()0,;,vv== y () ∑Y y() vmhv y y = Y y (), v ∈ Cc (20) n k=1 To verify the last condition we can solve the approximating problem defined by equation (20) kM∞, kM and determine the unknown parameters m y ∈ Cc , hy ∈ for all kn= 1,… , . Alternatively we can set

kk ∞,M Y yyyy()vm;, h= Y () v , v ∈ Cc , kn=1,… , (21)

In the latter case we must have a suitable kernel characteristic function such that (21) holds. This can be generally succeeded when the excitation process is described by a characteristic functional with explicit form (for example Gaussian or Gamma characteristic functional etc.). If this not the case we must determine the suitable parameters for the kernel characteristic functionals. In that case Theorem 4.3 guarantees that the equation (20) can be approximately satisfied. b2. Related with condition a3 for the initial state of the system we must have

nnυυ⋅=δδ,0 ⋅= YYxy()tt00 x () 1 n =⋅= υυδ ;,mhkk () ,υ ∈ N (22) ∑Y xyy()t0 Y 0 n k=1 As in the previous condition b1, similarly this condition will be satisfied by solving equation (22). The resulted values of the parameters will constitute the initial conditions for the set of differential equations that will be formulated using the Functional Differential Equation. Note that for the special case, where the initial states of the excitation and response are depending with each other the condition described above will take the form

n υν⋅⋅=δδ, Y xy ()tt00 1 n =⋅⋅ υνδδ,;,,,mmhhkkkk =() υν ,, υν∈∈NM, (23) ∑Y xxyxy()tt00 Y 0 n k=1 where Y 0 ()υν, will be the joint characteristic functional of the random variable ()xy()()0;ii , 0; : Ω→ NM × describing the initial state of the system.

188 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

Equations governing the parameters of the kernel characteristic functionals

We have determined the essential conditions for the parameters of the representation (19) so that conditions a1 and a3 are verified. We shall now derive an equivalent set of differential equations with the functional differential equation for the undetermined parameters of the kernel characteristic functionals. This set of equations will be formulated using a Gallerkin approach for the probability measures by direct application of kernel probability measures. In other words, if formally, we could write the linear operator equation for the probability measure of the solution as

⎡⎤xy,0= , for every xy∈∈CC∞∞,,NM, (24) L ⎣⎦⎢⎥c xy () cc then, using the kernel representation (18) and integrating equation (24) with respect to every probability measure we would have the equivalent set of equations

n ⎡ kj⎤ xy,,0,1, dd x y== j… n (25) ∑ ∫∫ L ⎢⎥ccxy() xy () k=1 ∞∞,,NM⎣⎦ CCcc×

Alternatively we could multiply with every kernel probability measure, instead of integrating. As a result we would the set of equations

n kj ⎡⎤ ∞∞,,NM ∑L ⎢⎥ccxy⋅== xy ()xy,0,1,j …n , for every xy∈∈CCcc, (26) k=1 ⎣⎦

Then we can integrate with respect to every measure and get a set of equations for the parameters of the kernel probability measures. This above method seems very promising, but the main problem is that in general (an important exception is the Fokker-Planck equation), we cannot have in an explicit form the operator L . Additionally it’s very difficult to operate with probability measures, since they are set functions. Instead of this, we have for a very wide class of dynamical systems an explicit form of the operator governing the characteristic functional. Such operators were studied, extensively in the present chapter at Section 5.2. Thus, we can express the dynamics of the system with an operator equation (related to (24) if exists) of the form

⎡⎤uv,0= , for every uv∈∈CC∞∞,,NM, (27) L ⎣⎦⎢⎥Y xy () cc

Moreover we have a wide class of explicit formulas for characteristic functionals as well as theorems for the analytical calculation of infinite dimensional integrals of those functionals. Hence, it would very interesting, if we could derive an equivalent set of equations with equations (25) or (26) using the characteristic functional approach. For simplicity we will first present the subject for the case of a probability measure defined on a one-dimensional space. Such a case can be met on a one dimensional dynamical system under independent increments excitation.

Hence the operator equation (24) will take the form of a partial differential equation for the probability density function

L[]()fx= 0 , x ∈ (28)

For the same case we would have partial differential equation for the characteristic function φ 5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 189

L[]()φ u = 0 , u ∈ (29)

Of course the following relation diagram holds,

F LL[fx ]()=∈0, x←⎯⎯⎯⎯ ⎯⎯→ []()φ u =∈ 0, u -1 F F fx(),, x∈∈←⎯⎯⎯⎯ ⎯⎯→ φ () u u F-1 where F denotes the Fourier transform operator. Now, consider the operator

∞ v2 1 − ⎡⎤ 2σ TL⎣⎦[]φψσ,; ()uuvvedvu=−⋅⋅∫ L []( φ ) ψ () , ∈ (30) −∞ σπ2 and φψ, are characteristic functions. Moreover the operator equation

⎡⎤ TL⎣⎦[]φψσ,; ()uu=∈∈ 0 , , ψ (31) can be easily verified that is equivalent with (29). We will study the behaviour of the above operator for extreme values of the parameter σ . For σ → 0 we will have

∞ v2 1 − limTL⎡⎤[]φψσ , ; ()uuvvedvu=−⋅⋅= L []( φ ) ψ () 2σ L []() φ , u ∈ σ→0 ⎣⎦∫ −∞ σπ2

Hence,

limTL⎡⎤[]φψσ , ; ()uu=∈⇔ 0 , L []() φ uu =∈ 0 , σ→0 ⎣⎦

Thus, for σ → 0 operator equation (31) interprets equation (24) in terms of the characteristic function. In contrast for σ →∞ we will have

∞ v2 1 − limσφψσσφ⋅=−⋅⋅=TL⎡⎤[] , ; ()uuvvedv L []( ) ψ () 2σ σ→∞ ⎣⎦∫ −∞ σπ2 ∞ =−⋅∫ L[](φψuv ) () vdv , u ∈ −∞ Hence,

limσφψσ⋅=∈⇔∗=∈TL⎡⎤[] , ; ()uu 0 , ⎡⎤ L [] φψ () uu 0 , (32) σ→∞ ⎣⎦ ⎣⎦

Thus, for σ → 0 we have (when operator L exists)

limσφψσ⋅=∈⇔⋅=∈TL⎡⎤[] , ; ()uu 0 , L [ fxgxx ]()() 0 , (33) σ→∞ ⎣⎦ where g =F −1 []ψ .

Now let us represent the solution of (28) using kernel representation f k

n 1 f ()xfx= ∑ k () (34) n k=1 190 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

Hence, the characteristic function will have the form

n 1 φφ()x = ∑ k ()x (35) n k=1 Applying the above representation into (33) we will have the set of equations

⎡⎤⎡⎤n ⎢⎥ limσφφσ⋅=∈=TL⎢⎥kl , ;()uuln 0 ,… , 1, , (36) σ→∞ ⎢⎥⎢⎥∑ ⎣⎦⎣⎦k=1

Because of (33) the above set will be equivalent wit the set of equations

n ⎡⎤ ∑L ⎢⎥f kl()xfx⋅=∈() 0, x… , l = 1, , n (37) k=1 ⎣⎦

Thus we have derived equations (36) that interpret equations (26) or (37) in terms of the characteristic function. We must note that for many kernel probability measures it’s not necessary to get σ →∞ to achieve equivalence of (31) with (37). More specifically for kernels characteristic functions that have compact support (like Gaussian measures) it’s sufficient to ⎛⎞v 2 ⎜ 0 ⎟ have a value of σ such that exp⎜−≈⎟ 1 where v0 is the radius of the support of the kernel ⎝⎠⎜ 2σ ⎟ characteristic function f (if exists). By inference we have achieved to derive an equivalent reformulation of the characteristic function differential equation (29) described by equation (31) which has the property to interpret (for special choice of the parameter σ ) equation (26) or (37).

Motivated from the above discussion we generalize for the general case where no explicit formula for the operator L exists. So, let us consider the general case of a system as described at the beginning of the present section (conditions a1, a2, a3) and let the associated functional differential equation described by the operator

⎡⎤uv,0= , uv∈∈CC∞∞,,NM, (38) L ⎣⎦⎢⎥Y xy () cc

Then, based on the analysis for the finite dimensional case described above we consider the operator

⎡⎤,;Cdduv ,=−−⋅⋅ u wv , z wz , w , z (39) TL⎣⎦[]YZ ( )∫∫ L []( Y )( Z ) cG ( ) ∞∞,,NM CCcc×

∞∞,,NM for every uv∈∈CCcc, , where the measure c G denotes a Gaussian measure into the ∞∞,,NM ∞∞ ,, NM measurable space ()CCcc××,U () CC cc with correlation operator C , centred at zero. When the norm of the correlation operator C tends to zero equation (39) will formally, interpret equation (24) and when the norm of the correlation operator C is sufficiently large equation (39) will formally tends to interpret the equation

⎡⎤xy,,0⋅= xy , for every xy∈∈CC∞∞,,NM, (40) L ⎣⎦⎢⎥cdxy ()() cc where d is the measure associated with the characteristic functional Z . Now consider the representation (19) satisfying conditions b1 and b2. Applying it at the operator equation

5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 191

⎡⎤ ,;C uv ,= 0 for every uv∈∈CC∞∞,,NM, (41) TL⎣⎦[]YZ ( ) cc

k 3 and setting ZY= xy ( ) for kn= 1,… , we will have the set of equations,

n jk ⎡⎤ ⎡⎤ ∞∞,,NM ∑TL⎢⎥⎢⎥YYxy,;,0 xy C ()uv = , uv∈∈CCcc, , kn=1,… , (42) j=1 ⎣⎦⎢⎥⎣⎦

We emphasize that all the infinite dimensional integration described above can be carried out analytically using the results of Section 5.4.1 of the present chapter. Hence we have achieved to interpret equation (26) in terms of the characteristic functional. We must emphasize that in the above discussion we don’t prove anything about equivalence between equations (40) and and (41) or between (42) and (26). We just argue using finite dimensional relatives that an efficient way to produce equations from the characteristic functional equation is by using operator (39). In this way achieve localization in the probability measure domain. Analogous comments concerning the correlation operator C can be made as in the case of the finite dimensional case. Thus if we choose the correlation operator such that the rate of decay ∞∞,,NM of probability for the elements wz∈∈CCcc, is less that the rate of decay of the values ∞∞,,NM of the characteristic functionals [](YZuwvz−−⋅,, )( wz ) for every uv∈∈CCcc, then we would have achieved the desired result. Finally since the equations (42) expresses the desired properties for the probability measures ∞∞,,NM we can either integrate with respect to a Gaussian measure over uv∈∈CCcc, analytically (again we choose a zero centred Gaussian measure with suitable correlation operator as before), or we can take moment equations for every equation (42) using the results of Section 5.3.2 of the present chapter. We must also notice that another approach for the derivation of simpler set of equations from the functional differential equation is based on the direct integration of the equation (38) with respect to various probability measures. Is should be emphasized that since we relate characteristic functionals with probability measures the choice of the probability measures should be very careful if we want to maintain all the necessary information including into equation (38). The above it will not be studied into the present work. Now let us illustrate the above discussion by studying a specific system

Example 4.16 : To describe the proposed method we will study the specific case of a first order stochastic differential equation with cubic nonlinearity, i.e. the system

x()taxtyt;;;βββ= ()3 + () a.s. (43)

xt()()00;ββ= x , a.s. (44)

All the involved functions are assumed to be real valued. We have seen that the associated functional differential equation has the form

k 3 k kkk ( ) The notation Y xy for Y xy()uvm,; x , m y , h x , h y will also be used if no confusion is likely to occur. 192 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

d ⎡⎤δδδYYY()uv,,,3 () uv() uv ⎢⎥xy=−a xy + xy (45) ⎢⎥ 3 dt⎣⎦⎢⎥δδ u() tδut() v() t

YYxy ()υδ00,0=∈ () υ , υ (46) where we have assumed that the same conditions a1, a2 ,a3 mentioned before, are fulfilled. Consider the representation of the joint characteristic functional for the response and the excitation

n n1 k kkk YYxyxyxyxy()uv,,;,,,= ∑ () uvm m h h (47) n k=1 where kkkk a) Y xyxyxy()uvmmhh,; , , , ()kn=1,… are the characteristic functionals corresponding to k kkk the kernel probability measures c xyxyxy(x,;ym , m , h , h) as described at Definition 4.7.

b) The characteristic functional is compatible with the characteristic functional of the n nn1 kk excitation process, i.e., YYxyy()0,vvpvmh== () ∑ kyyy Y() ; , . Hence the n k=1 kk parameters mhyy, ()kn=1,… are given.

c) The characteristic functional satisfies the initial conditions (44) or (46), i.e., n nn1 kk YYxyx()υδ()ii,0==() υδ() ∑ Y yxx() υδ() i ;mh , (48) n k=1

Substituting representation (47) into the functional differential equation (45) we have

kkk nnn 3 d δδδYYYxy()uv,,, xy() uv xy () uv ∑∑∑=−a 3 + (49) dtkkk===111δδ u() tδut() v() t

l Multiplying equation (49) with Y xy ()wuzv−−, and integrating over a Gaussian measure with zero mean value and covariance operator C , with respect to uv , we have the set of ∞∞ equations ()ln=1,… for every uC∈∈cc , vC

k n d δY xy ()wz, l u−− w,, v z dw dz = ∑ ∫ Ycxy()() G dtk=1 ∞∞ δ w() t CCcc× k 3 n δ Y ()wz, l =−auwvzdwdzxy −,, − + (50) ∑ ∫ 3 Ycxy()() G k=1 ∞∞ δwt() CCcc× k n δY xy ()wz, l +−− u w,, v z dw dz ∑ ∫ Ycxy()() G k=1 ∞∞ δzt() CCcc×

Using the notation

uw−= u dw⎡ Cu⎤ ∫ YYcYY12( ) ()G ( )T ⎣⎢ 12,,⎦⎥ () ∞∞ CCcc×

∞∞ we have the set of equation ()ln=1,… for every uC∈∈cc , vC 5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 193

kk nn⎡⎤ ⎡3 ⎤ δδll d ⎢⎥YYxy ⎢ xy ⎥ ∑∑TT⎢⎥,,YYxyCuv(),,,,=− a ⎢3 xy Cuv ⎥() + dtkk==11⎢⎥δ u() t ⎢δut() ⎥ ⎣⎦⎢⎥ ⎣⎢ ⎦⎥ (51) k n ⎡⎤ δ l ⎢⎥Y xy + ∑T ⎢⎥,,Y xy Cuv(), ⎢⎥δvt() k=1 ⎣⎦⎢⎥

The above set of equations formulates a set of functional differential equations for the parameters of the kernel characteristic functionals with initial conditions defined from equation (48). The difference with the prior functional differential equation (49), is that every equation of the last set, controls a specific part of the probability space and is coupled with the other parts through the coupling of the functional differential equation set (51). Since the integrals are over Gaussian measures and the integrated functionals involve linear and quadratic forms, Theorems 4.1 and 4.4 can be applied for the analytical calculation. We will not pass into details for the analytical calculation of the above integrals. The resulted set of functional equations can be numerically solved by either integrate them analytically with respect to a suitable Gaussian measure or by getting moment equations using the result of Section 5.3.2 of the present chapter. In the following section we will try to simplify further the set of functional equations (42) and get numerical results without even make the analytical calculations of the functional integrals. In the next section a special simplification will be applied to equations (50) to get numerical results without having to calculate the above integrals. ■ 194 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

5.4.5. Simplification of the set of FDE’s governing the Kernel Characteristic Functionals.

In this section we shall simplify the set of functional differential equations (42) to obtain a set of ordinary differential equations for immediate numerical solution. Hence, based on the following simplification we will not need to evaluate analytically the functional integrals. However we must emphasize, that in this way, we just illustrate the validity of the proposed method. We don’t expect to have the best possible numerical results. For this reason the following analysis will mainly concern the first order system with cubic nonlinearity, described at the previous sections.

x()tkxtaxtyt++ () ()3 = () (52)

Analogous arguments can be used for the analysis of higher dimensional systems with more complex polynomial nonlinearities. For the numerical implementation we will use Gaussian kernel characteristic functionals.

The simplification will be based on the restriction of the concentration parameter (for Gaussian kernel characteristic functionals this parameter will be the variance) to sufficiently small values such that the interaction between every pair of Gaussian kernels could be considered negligible. However this assumption cannot hold for a great time since the equations governing every kernel leads to greater values of the concentration parameter. Thus, to avoid growth of the concentration parameter we approximate, when a variance exceed a certain value, the current probability measure by another set of kernel measures with different amplitudes and smaller variances. Then this set of kernel measures evolutes with respect to the dynamical equations until some concentration parameter exceed a certain value. The accuracy of the proposed method mainly depends on the threshold value of the concentration parameter.

Hence we use a kernel characteristic functional representation

n n YY()ii= ∑ pkxkykk();,,mmh,, , with pn ≥ 0 , nN= 1,… , , ∑ pn = 1 (53) k=1 n

where every Y (i;,,mmhx,,kykk) ()kn=1,… , is a Gaussian kernel characteristic functional. Based on the above assumptions we can conclude that the set of equations (42) will take the form

⎡⎤⎡⎤kk TL⎢⎥ ⎢⎥YY,;,0C ()uv = , , kn= 1,… , (54) ⎣⎦⎢⎥⎣⎦xy xy

or,

k ⎡⎤ ∞∞,,NM L⎢⎥Y ()uv,0= , uv∈∈CC, , kn=1,… , (55) ⎣⎦xy cc n Note that for the case described we need general amplitudes 0≤≤pk 1, ∑ pk =1 (instead of k=1 equal with each other, assumed at the previous sections) for every kernel characteristic functional. These amplitudes will be assigned from the approximation algorithm, every time that a concentration parameter exceeds the threshold value. Now let us derive the ordinary 5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 195

differential equations associated with every Gaussian kernel, for the specific case of the one dimensional system with a linear and cubic term. Equations (55) will take the form

kk kk 3 d δδδδYYxy xy YY xy xy ()uv,,+− k() uv a() uv ,, =() uv (56) dtδδ u() t u() tδut()3 δ v() t

∞∞ for every uC∈∈cc, vC()kn=1,… . We observe that every characteristic functional controls a specific region of the probability space and evolutes independently from the others. Note also, that the interchange of probability between the kernels characteristic functionals is carried on completely from the approximation algorithm and only at the case when some concentration parameter exceeds the threshold value. Although this assumption is very loose, since the probability must interchange continuously (with respect to time), between the kernels, it gives us a first picture of the method. We emphasize that with the analytical calculation of the functional integrals the interchange of probability occurs continuously since the system of ordinary differential equations governing the kernel parameters is coupled. Hence, for this case the interchange of probability is introduced from the nondiagonal terms of equation (42).

An important feature of the discussed method is the ability for parallelized computations. Parallelization techniques can be physically applied both to the dynamical evolution of the kernels and the optimization algorithm. For the first case the algorithm can take advantage of the independent evolution of every kernel. Additionally for the optimization algorithm we can approximate the given density (since is already splitted into Gaussian kernels) by independent group of kernels with small variance. Hence, in this way we can succeed very fast computations for system of higher dimensions, with various excitations.

To derive a set of ordinary differential equations for every FDE (56) we will get moment equations using the result of Section 3 of the present chapter. Hence, since every characteristic k functional Y xy is Gaussian, the functional differential equation (56) will be fulfilled if the three first moment equations are satisfied. By direct application of Section 5.3.2 of the present Chapter and Appendix I lead us to the following subset of ordinary differential equations for every kernel ()kn=1,… 3 mkk t++ kmt3, amtBttamt kkkk +⎡ ⎤ = mt Xt,,() X () X ,,,, () XX ( )⎣⎢ X ()⎦⎥ Y ()

kk kkkkkk BXX,,,,,,,, tt()ts,,3,3,,,++ kB XX () ts am X ()() tB XX tsm X () t + aB XX () ttB XX () ts = B XY () st

kk kkkkkk BXY,, tt()ts,,3,,3,++ kB XY () ts aB XX ,,,,,, () ttB XY () ts + am X ()() tB XY tsm X () t = B YY () ts , for all (ts, )∈× I I. We will not come into details concerning the numerical solution of the above equations. We just emphasize that special treatment was needed, since as we note, there is no participation of the s variable to the dynamics of the equations. Hence we have to solve the above system for every value of s. Let us now give a precise description of the optimization algorithm used for the rebuild of a measure (more precisely density) with kernels with smaller variance. As we mention we will have to approximate a given superposition of Gaussian densities by another superposition with the restriction of a maximum variance. We thus have the following optimization problem, 196 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

2 ⎛⎞22 ⎜+∞ ⎡⎤NM ⎪⎪⎧⎫⎛⎞ ⎪⎧ ⎛ ⎞ ⎪⎫⎟ ⎜ ⎢⎥Axm⎪⎪11− Axmjj ⎪− ⎪⎟ min⎜ ii exp⎪⎪−−⎜⎜⎜⎜⎟⎟⎟ exp ⎪ − ⎪dx⎟ M ⎜ ⎢⎥∑∑⎨⎬⎟⎟ ⎨ ⎬ Am,,σ ⎜∫ ⎜⎜⎟⎟⎟ {}ii0 i=1 ⎢⎥σπ 22⎪⎪22⎝⎠⎜⎜σσ σπ ⎪ ⎝ ⎠ ⎪⎟ ⎝⎠⎜−∞ ⎣⎦ij==11i ⎩⎭⎪⎪i 0 ⎩⎪0 ⎭⎪⎟ N where Am ,, σ are the amplitudes, mean values, variances of the given superposed {}iiii=1 kernels and Am,,σ M are the parameters of the kernels that approximates the given {}ii0 i=1 density. We also restrict ourselves, for simplicity, to the case of a common variance σ0 for all the Gaussian kernels. More details for the optimization algorithm can be found at Appendix II. In what follows we will examine the case of the above system with two different set of parameters. Hence a white noise excitation case will be studied with the system parameters described by the table

Non-dimensional Case I Case II parameters k 1 1 a 1 -1 White noise intensity 1 1

We note that for the first case the nonlinear system has one stable fixed point located at zero. As the parameter a decreases a pitchfork bifurcation occurs at zero and two symmetric stable k points appear at ± and the fixed point at zero becomes unstable. Hence we have the a bifurcation diagram

Bifurcation Diagram 1.5

1

0.5

0 Fixed point x

-0.5

-1

-1.5 -1.5 -1 -0.5 0 0.5 1 1.5 Bifurcation parameter

The above analysis will also be confirmed from the resulted densities. Thus we expect that for the case II, the probability will concentrate around the fixed points and hence we will have a bimodal distribution. For the first case we expect a unimodal distribution with the probability concentrated at zero. Since the above fixed points are global attractors we expect to have these 5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 197

results after some time, independent of the initial density. For the numerical results presented above we are using two cases of initial distributions. The first case is consisting of a Gaussian distribution cantered at zero with varianceσ = 0.1. The second case is a convex superposition of two Gaussian distributions with characteristics mm12==0, 0.6,σσ 1 = 0.1, 2 = 0.6 and amplitudes a1 = 0.4 and a2 = 0.6 respectively.

f tx, for the case of Gaussian White Noise excitation with one stable fixed point at 0 and initial conditions of the first type. x ()

The first plot shows with solid lines the mean values of the kernel probability measures. Their shape verifies the stable property of the fixed point at 0. We can see that the initial distribution is spreading and approaches very the final steady state distribution. A very interesting point is that the concentration of probability is not stopping as can be seen from the outer kernels. This behaviour reveals the necessity for the study of the tail behaviour, since a representation which predicts the tail revolution will overcome the mentioned problem. 198 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

f tx, for the case of Gaussian White Noise excitation with one stable fixed point at 0 and initial conditions of the second type. x ()

Again we can see from the first plot that the kernels mean values are attracted from the stable fixed point (global attractor) In this case, since the initial distribution is more complex more kernels are used for the efficiently representation of the solution. Notice that the interchange of probability between kernels takes place every 0.2 seconds, which is every time we have discontinuity of the mean value curves.

Notice that since the steady state probability distribution has an asymptotic behaviour of 4 oe()−x and simultaneously we are using Gaussian representations, we will constantly have the previously described behaviour of the outer kernels. Hence, it is very important to use suitable kernel representations with identical tail behaviour. However, we can see that the above representation produces satisfactory results for the main probability mass.

5.4. KERNEL REPRESENTATIONS FOR THE CHARACTERISTIC FUNCTIONAL 199

f tx, for the case of Gaussian White Noise excitation with two stable fixed points at ±1 and initial conditions of the first type. x ()

In this case we examine the behaviour of the system for different system parameters. Hence, we have for the underlying Hamiltonian system two stable fixed points at ±1 while the fixed point at 0 has become unstable. For the first case of initial conditions we have full symmetry with respect to the spatial 0. The symmetry is maintained for the kernel mean values as well. Note also that, despite the discontinuities of the kernels mean values the probability distribution is continuous (second plot).

After a short time we can see that the probability is concentrated near the stable fixed points and a steady state is approached. Again we can see the higher concentration of outer kernels, something that can be explained from the tail evolution problems. It must also be noted that the rapid spread of probability is explained from the chosen white noise intensity.

200 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

f tx, for the case of Gaussian White Noise excitation with two stable fixed points at ±1 and initial conditions of the second type. x ()

This is the most interest case. We start with a bimodal initial condition and we are approaching a bimodal steady state distribution. From the first plot we note that until the time of 0.4 seconds the starting bimodal distribution is transforming to a unimodal one. As the system evolves, the probability mass is transferred from the stable fixed point located at -1 to the stable fixed point at 1. This is a very interesting phenomenon since between the two stable fixed points is an unstable one which divides the phase space of the deterministic problem and does not allow the transition of orbits from the one sine to another. This is not the case for the probability mass, were we know that after a sufficient time period the probability of the two stable fixed points regions will be equal.

5.5. REFERENCES 201

5.5. References

ATHANASSOULIS, G.A. & GAVRILIADIS, P.N., 2002, The Truncated Hausdorff Problem solved by using Kernel Density Functions. Probabilistic Engineering Mechanics, 17, 273-291. BATCHELOR, G.K., 1953, Theory of Homogeneous Turbulence. Cambridge University Press. BERAN M.J., 1968, Statistical Continuum Mechanics. Interscience Publishers. CARTAN, H., 1971, Differential Calculus. Kershaw Publishing. DI PAOLA, M. & SOFI, A., 2002, Approximate Solution of the Fokker-Planck-Kolmogorov Equation. Probabilistic Engineering Mechanics, 17, 369-384. EGOROV, A.D. & SOBOLEVSKY, P.I. & YANOVICH, L.A., 1993, Functional Integrals: Approximate Evaluation and Applications. Kluwer Academic Publishers. FALKOVICH, G. & LEBEDEV, V., 1997, Single Point Velocity Distribution in Turbulence. Physical Review Letters, 79, 4159-4161. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1996, Theory of Random Processes. Dover Publications. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1974, The Theory of Stochastic Processes I. Spinger. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1975, The Theory of Stochastic Processes II. Spinger. GIΚHMAN, I.I. & SKOROΚHOD, A.V., 1975, The Theory of Stochastic Processes III. Spinger. GUO-KANG ER., 2000, Exponential Closure Method for some Randomly Excited Non-Linear Systems. International Journal of Non-Linear Mechanics, 35, 69-78. HOPF, E., 1952, Statistical Hydromechanics and Functional Calculus. Journal of Rational Mechanics and Analysis 1, 87-123. JOE, H., 1997, Multivariate Models and Dependence Concepts, Chapman & Hall, London. KONSTANT, D.G. & PITERBARG, V.I., 1993, Extreme values of the cyclostationary gaussian random process. Journal of Applied Probability. KOTULSKI, Z. & SOBCZYK, K., 1988, On the Moment Stability of Vibratory Systems with Random Parametric Impulsive Excitation. Arch. Mech., 40,4, 465-475. LANGOUCHE, F. & ROEKAERTS, D. & TIRAPEGUI, E., 1979, Functional Integral Methods for Stochastic Fields. Physica, 95A, 252-274. MONIN. A.S. & YAGLOM, A.M., 1965, Statistical Fluid Mechanics: Mechanics of Turbulence. MIT Press. MOSBAH, H. & FOGLI, M., 2003, An Original Approximate Method for Estimating the Invariant Probability Distribution of a Large Class of Multi-Dimensional Nonlinear Stochastic Oscillators. Probabilistic Engineering Mechanics, 18, 165-170. POLIDORI, D.C. & BECK, J.L. & PAPADIMITRIOU, C., 2000, A New Stationary PDF Approximation for Non-Linear Oscillators. International Journal of Non-Linear Mechanics, 35, 657-673. PRADLWARTER, H.J., 2001, Non-Linear Stochastic Response Distributions by Local Statistical Linearization. International Journal of Non-Linear Mechanics, 36, 1135-1151. PROHOROV, YU.V., 1956, Convergence of Random Processes and Limit Theorems in Probability Theory. Theory of Probability and Its Applications, I, 157-214. PROHOROV, YU.V. & ROZANOV, YU.A., 1969, Probability Theory. Springer. 202 CHAPTER 5 PROBABILISTIC ANALYSIS OF STOCHASTIC DYNAMICAL SYSTEMS

PROHOROV, YU.V. & SAZONOV, V.V., 1961, Some Results Associated with Bochner’s Thoerem. Theory of Probability and Its Applications, VI, 82-87. PUGACHEV, V.S. & SINITSYN , 1987, Stochastic Differential Systems. John Wiley &Sons. ROSEN, G., 1967, Functional Integration Theory for Incompressible Fluid Turbulence. The Physics of Fluids, 10, 2614-2619. ROSEN, G. & OKOLOWSKI, J.A. & ECKSTUT, G., 1969, Functional Integration Theory for Incompressible Fluid Turbulence. II. Journal of Mathematical Physics, 10, 415-421. ROSEN, G., 1960, Turbulence Theory and Functional Integration I. The Physics of Fluids, 3, 519-524. ROSEN, G., 1960, Turbulence Theory and Functional Integration II. The Physics of Fluids, 3, 525-528. SKOROKHOD, A.V., 1974, Integration in Hilbert Spaces. Spinger. SKOROKHOD, A.V., 1984, Random Linear Operators. D. Reidel Publishing Company. SOBCZYK, K., 1991, Stochastic Differential Equations. Kluwer Academic Publishers. SOONG, T.T., 1973, Random Differential Equations. Academic Press. TATARSKII, V.I., 1995, Characteristic Functionals for one Class of Non-Gaussian random Functions. Waves in Random Media, 5, 243-252. VISHIK, M.J. AND FURSIKOV, A.V., 1980, Mathematical Problems of Statistical Hydromechanics. Kluwer Academic Publishers.

Subjects for Future Research

As it has been seen in the text there are numerous threads for future investigations. One very interesting direction would be the application of the results presented for the numerical solution of stochastic differential equations describing complex dynamical systems. Some topics on this direction of future research should be • Implementation of numerical method for the solution of the partial differential equations derived at Section 5.3.4 through the reduction of the Functional differential equation. In this way we would have results for the joint characteristic function of the response, its derivatives and the excitation at various time instants. Probably the most convenient way should be the use of kernel characteristic functions representations with given marginals. • Direct application of the proposed method of Section 5.4.4 without any simplifications. In this way we could solve the full statistical problem and get the complete probability structure through the characteristic functional. This direction should require analytical computation of infinite dimensional integrals by applying the theorems presented at Section 5.4.1. • Calculation of the steady-state probability measures associated with a dynamical system, by means of suitable kernel characteristic functional representation. This representation should describe a stationary measure and its application for the numerical computation of its parameters might use a variation of the proposed method of Kernel Characteristic Functionals. • Improvement of the simplified method, presented at Section 5.4.5, and especially of the optimization algorithm described at Appendix III for faster computation of the probabibility densities as well as for numerical solution of higher dimensional problems. In this direction should be take in account the embedded characteristic of the method for parallel computation, discussed at Section 5.4.5. Another research thread of great importance should be the analytical investigation of dynamical systems by means of the kernel probability measure representations. More specifically some topics on this direction could be • Study of the evolution of the asymptotic behaviour (tail) of the response, directly thorough the analytical study of the characteristic functional and more specifically thought the study of the asymptotic behaviour of the characteristic functional at the neighbourhood of zero. In this direction a useful tool might be the Tatarskii characteristic functional. The above topic is of great importance since the knowledge of the tail behaviour will allow solving the problem with few kernels. • Analytical study of the behaviour of the probability structure of the response by studying analytically the parameters of the kernel characteristic functionals in relation with critical system parameters. An interesting point on this would be how the bifurcation analysis of the underlying deterministic system would/could predict the probability structure of the response through the kernels. In this way we could study more clearly the effects of nonlinearity and stochasticity simultaneously.

203 204 SUBJECTS FOR FUTURE RESEARCH

Finally an important topic concerning the study of kernel characteristic functionals is the proof of theorems analogous to theorem 4.13 of Section 5.4.3 due to D.G. Konstant and V.I. Piterbarg, concerning the probability structure of the extremes of stochastic processes with known probability measure. Hence in this way we would have necessary conditions for a probability measure/characteristic functional to be admissible for the solution of the statistical problems.

Appendix I

Volterra Derivatives of Gaussian Characteristic Functional

For the sake of convenience we shall now compute, the essential derivatives of Gaussian functional for our analysis. Let, ⎛⎞ ⎜ 11 ⎟ Y ()uv,exp,=+−−−⎜ imX u im Y , v C XX uu , C YY vv , C XY uv ,⎟ ⎝⎠⎜ 22 ⎟ Then we have the Volterra derivatives computed at ()0,0 .

δY ()0,0 1 = mt() δuiX 1 t1

δY ()0,0 1 = mt() δviY 1 t1

δ 2Y ()0,0 1 =+B ()tt, m ()() t m t δδuu i2 XX12 X 1 X 2 tt12

δ 2Y ()0,0 1 =+B ()tt, m ()() t m t δδuv i2 XY12 X 1 Y 2 tt12

δ 2Y ()0,0 1 =+B ()tt, mtmt ()() δδvv i2 YY12 Y 1 Y 2 tt12

δ3Y ()0, 0 1 =++mtB() ( tt,,, ) mtB () ( tt ) mtB () ( tt ) δδδuuui3 XXXX123213312 XXXXX tt123 t

+ mtmtmtXX()123 () X ()

δ3Y ()0, 0 1 =++mtB() ( tt,,, ) mtB () ( tt ) mtB () ( tt ) δδδuuvi3 XXYX123213312 XYYXX ttt123

+ mtmtmtXXY()123 () ()

205 206 APPENDIX I

δ 4Y ()0,0 1 =++BttBttBttBtt()(,, ) ()() ,, δδδδuuuui4 XX13 XX 24 XX 23 XX 14 tttt1234

+++mtBXXXX()1 ( ttmt 2342,, ) () mtB X () XXX ( ttmt 134 ) ()

+++mtBXXXX()1 ( ttmt 2432,, ) () mtB X () XXX ( ttmt 143 ) ()

++mXXXX() tB3124 ( ttmt, ) () m XX () tmtm 1234 () X () tm X () t +

++BttBttmtBttmtXX()()12,, XX 34 X ()()() 1 XX 34 , X 2 δ 4Y (0,0) 1 =++B()( ttB,, tt ) B ()() ttB ,, tt δδδδuuuvi4 XX13 XY 24 XX 23 XY 14 tt1234 tt

+++mtBXXXY()1 ( ttmt 2342,, ) () mtB X () XXY ( ttmt 134 ) ()

+++mtBXXYX()1 ( ttmt 2432,, ) () mtB X () XYX ( ttmt 143 ) ()

++mXXXY() tB3124 ( ttmt, ) () m XX () tmtm 1234 () XY () tmt () +

++BttBttmtBttmtXX( 12,,) XY( 34) X( 1) XY( 34 ,) X ( 2)

The above calculations will for the derivation of moment equations from the FDE. These equations correspond to the Gaussian closure technique. In Section 5.4.5 of the Chapter 5 they will be used for the application of the Gaussian closure technique locally in the phase space and will allows us to get a first picture of the proposed method. Appendix II

Optimization Algorithm

We have the following optimization problem,

2 ⎛⎞22 ⎜+∞ ⎡⎤NM ⎪⎪⎧⎫⎛⎞ ⎪⎧ ⎛ ⎞ ⎪⎫⎟ ⎜ ⎢⎥Axm⎪⎪11− Axmjj ⎪− ⎪⎟ min⎜ ii exp⎪⎪−−⎜⎜⎟⎜⎜⎟⎟ exp ⎪ − ⎪dx⎟ M ⎜ ⎢⎥∑∑⎨⎬⎟⎟ ⎨ ⎬ Am,,σ ⎜∫ ⎜⎜⎟⎟⎟ {}ii0 i=1 ⎢⎥σπ 22⎪⎪22⎝⎠⎜⎜σσ σπ ⎪ ⎝ ⎠ ⎪⎟ ⎝⎠⎜−∞ ⎣⎦ij==11i ⎩⎭⎪⎪i 0 ⎩⎪0 ⎭⎪⎟ N where Am ,, σ are the amplitudes, mean values, variances of the given superposed {}iiii=1 kernels and Am,,σ M are the parameters of the kernels that approximates the given {}ii0 i=1 density. We also restrict ourselves, for simplicity, to the case of a common variance σ0 for all the Gaussian kernels. Hence we will have the restrictions, M LB<<σ0 UB and ∑ Aj =1 j=1 Expanding the above, we have the function for minimization

+∞ ⎧⎫2 2 NN AA ⎪⎪⎛⎞⎛⎞ x− m ij⎪⎪11xm− i ⎟ ⎜ j⎟ FA(){},,σ M=−−− exp⎨⎬⎪⎪⎜ ⎟ ⎜ ⎟ dx i 0 ∫ ∑∑ ⎪⎪⎜ ⎟ ⎜ ⎟ ij==11σσij22 π⎝⎠ σ i 2⎝⎠⎜ σ j ⎟ −∞ ⎩⎭⎪⎪ 22 +∞ NM AA ⎪⎪⎧⎫⎛⎞⎛⎞ x− m ij⎪⎪11⎜⎜xm− i ⎟⎟ j −−−+2exp∑∑ ⎨⎬⎜⎜⎟⎟dx ∫ σσ22 π⎪⎪⎝⎠⎝⎠⎜⎜ σ⎟⎟ 2 σ −∞ ij==11ii00⎩⎭⎪⎪ 22 +∞ MM AA⎪⎪⎧⎫⎛⎞⎛⎞ x− m ij⎪⎪11⎜⎜xm− i ⎟⎟ j +−−∑∑ exp⎨⎬⎜⎜⎟⎟dx . ∫ σσ22 π⎪⎪⎝⎠⎝⎠⎜⎜ σ⎟⎟ 2 σ −∞ ij==1100⎩⎭⎪⎪ 0 0 The integrations can be carried out analytically. Hence we will have, ⎧⎫2 NN ⎪⎪mm− AAij ⎪⎪()ji FA(){}i ,,σ0 M=−− exp⎨⎬ ∑∑ 22 ⎪⎪22 ij==112 σσπ+ ⎪⎪σσji+ ()ij ⎩⎭⎪⎪ 2 MN ⎪⎪⎧⎫ AA ⎪⎪()mmji− −−+2expij ⎨⎬⎪⎪ ∑∑ 22 ⎪⎪22 ij==112 σσπ + ⎪⎪σσ0 + i ()i 0 ⎩⎭⎪⎪ 2 MM ⎪⎪⎧⎫ AA ⎪⎪()mmji− +−ij exp⎨⎬⎪⎪. ∑∑ 22 ⎪⎪22 ij==112 σσπ+ ⎪⎪σσ00+ ()00 ⎩⎭⎪⎪

To minimize the above quantity we fixed the common variance σ0 and the mean values

mii ,1,2,= … M defined by

⎡⎤31 =−⎢⎥σσ +⎡ −+ ⎤ mmi min()00 max() ⎣ max() mm 000 min () 3max ()⎦ ⎣⎦⎢⎥21M −

207 208 APPENDIX II

Thus the minimization problem is equivalent with the set of equations,

BijAC j= i where, ⎪⎪⎧⎫2 1 ⎪⎪()mmji− B =−exp⎨⎬⎪⎪ iMjM==1,…… , 1, , ij ⎪⎪2 22σπ0 ⎪⎪2σ0 ⎩⎭⎪⎪ N ⎪⎪⎧⎫2 A ⎪⎪()mmki− =−k ⎪⎪= … Ci ∑ exp⎨⎬2 iM1, 22σπ ⎪⎪2σ k=1 0 ⎩⎭⎪⎪0

And the value of the minimizing function will be,

FA(){}i,,σ0 M=−+ ADACAABA i ij j 2 i i i ij j

To the linear system of equations we must add the condition M ∑ Aj =1 j=1 Thus we have a least-square problem, with M +1 equations. For our case we make the function

GMFA()σσ0,0,,,= (){}iopt M with the amplitudes {}Aiopt, defined by the least-square problem. Thus, for every M we minimize the function GM()σ0 , with respect to σ0 and then we choose the optimum

σ0,opt,,MA opt{} i , opt .

Appendix III

Application on Ship Stability and Capsizing on Random Sea

The Probabilistic Model

The current regulations and criteria for assuring the stability of a ship and preventing it from capsizing (see the Codes of the International Maritime Organization (IMO) and of the Germanischer Lloyd) are empirical and are based on the properties of the righting lever of the ship, taking only hydrostatic forces into account. For details and criticism see Kreuzer and Wendt (p. 1836). Those static criteria which neglect the motion of the ship as well as the impact of seaway and wind do obviously not guarantee total stability as they cannot prevent at least 30 ships yearly of tonnage greater than 500 GT from getting lost due to severe weather conditions. Hence researchers agree that those criteria have to be modified by using hydrodynamic models of the ship–sea system, by describing the sea as a random field and by analyzing the ship as a rigid body with 6 degrees of freedom using methods of nonlinear dynamics and the theory of random dynamical systems.

The present state of (deterministic as well as stochastic) ship dynamics research has been recently documented in the Theme Issue of the Phil. Trans. Royal Soc. London (Series A) entitled “The nonlinear dynamics of ships”, edited by Spyrou and Thompson. The volume includes the extended overview by Spyrou and Thompson, presenting also the historical development of the field as well as future directions of research, the prime one being the “integration of the probabilistic character of the seaway into the study of nonlinear dynamics” – which we will try to address.

In the following analysis we will not consider the linear model for rolling equation since it is the result of the assumptions that the wave is small. Hence the linearized model describes small ship motions. Transition to large roll angles requires the introduction of nonlinear terms into the ship motion equations. Unfortunately, there is no theory available to derive ship motions equations in a consistent manner. So we used a physics based approach, considering forces of a different nature separately and trying to combine them into one model of nonlinear rolling, suitable for capsizing study.

The forces acting on a ship in a seaway can be classified in the following manner BELENKY L.V. AND SEVASIANOV N.B., (Stability and Safety of Ships): • Inertial hydrodynamic forces (e.g. added mass) • Wave damping hydrodynamic forces • Viscous damping forces • Hydrostatic forces • Wave excitation forces • Other forces, including

209 APPLICATION ON SHIP STABILITY AND CAPSIZING ON RANDOM SEA 210

ƒ Aerodynamic forces ƒ Appendage forces (including dumping form rollers and anti-rolling fins) ƒ Forces due to fluid motion in internal tanks including anti-roll tanks ƒ Hull “manoeuvring” forces, propeller thrust, etc.

There are other systems to classify these forces. This classification assigns forces of a different nature to the same category. However, it reflects a role that force plays in ship motions.

For our analysis we will use the simplified, but nonlinear, rolling equation

()I xx+++⋅=⋅⋅a44 φφ2 Nφ mg GZ ()() φ mg GM a E t (1) where • We have neglect nonlinear damping forces • We have neglect the effect of bulwark entering into water during heeling • The change of GZ curve due to irregular seas • we have neglect effects of special possibly anti-rolling devices

• Nφ is a linear dumping coefficient

• atE ( ) describes the excitation due to wave forces

To “randomize” a capsizing model we need to : • Introduce a model for the stochastic excitation due to wave forces • Calculate probabilistic characteristics of rolling, roll velocities and other processes included in the model of capsizing • Calculate probabilistic characteristics of the “carrier process” – the process, by which crossing of the boundary is associated with the capsizing • Calculate the probability of crossing

In the direction described above we will express the stochastic forces using the well-known model of Pierson – Longuet Higgins Model (see Chapter 2). Hence we assume that atE ( ) is a normal stationery stochastic process with given spectrum or correlation function. Moreover we assume that all the necessary ship’s characteristics (such as GZ curve, GM, etc.) are known.

For the probabilistic analysis we will first write eq. (1) in a non-dimensional, state-space form, i.e.

φθ = (2) θθ++2bV′() φ = FtW () where,

• bN=+φ () Ixx a44 mg • VGZd()φϕϕ= ∫ () ()Iaxx + 44

• FWExx() t=⋅ mg GM ⋅ a ()( t I + a44 )

211 APPENDIX III

Functional Differential Equation for the Rolling Equation

We shall derive the corresponding functional differential equation for the stochastic system (2). We assume that

a) FIW ()ii;:×Ω → is a normal, stationary, mean square continuous stochastic process defined onto the probability space ()ΩΩ,,Uc() with the associated probability

∞∞ space ()CCcc,,Uc()F , and with characteristic functional Y F .

2 b) ()θφ00()ii,: () Ω→ are random variables with the associated probability space

22 () ,,Uc() θφ and characteristic functions Y θφ . 00 00

c) VI′()i: ×→ is a measurable, deterministic function that can be expanded into a series of the form

3 Vaa′()φφφ=+13 (3)

Applying Theorem 2.1 of Chapter 5 we will have the functional differential equation

d δδYYφθφθFF()uuv,, φθφθ() uuv ,, = (4a) dtδδ uφθ() t u() t

d δδδYYYφθFFF()uuv φ,, θ φθ() uuv φ ,, θ φθ() uuv φ ,, θ +++2ba1 dtδδδ utθθφ() ut() ut() (4b) 3 δδYYφθφθFF()uuv,, φθφθ() uuv ,, −=a 3 3 δvt δutφ () ()

Where Y φθF ()uuv φ,, θ is the joint characteristic functional of the stochastic processes FW ,φθ , . Concerning the initial conditions of the problem we have

υδ()ii,,0 υδ () = Y φθFt() φ00 θ t =++=expiiiFdddFφυδ ,()ii θυδ ,() ,0() φ , θ , (6) ∫∫ {}φθtt00 c φθ F ∞∞ CCcc =+expitυφ() υθ () t() dddF φ , θ , = υ , υ ∫ ()φθφθ00cYF θφφθ00() ∞ Cc

APPLICATION ON SHIP STABILITY AND CAPSIZING ON RANDOM SEA 212

PDE for the joint probability density function of the r.v.’s FtW (),,φθ () t () t

Following the analysis of Chapter 5/Section 5.3.4 we have the reduction of the Functional Differential Equation (4a) and (4b) to the PDE

∂∂∂∂φφφφφθFFFF()tts,, φθ() ttt ,, φθ() ttt ,, φθ () ttt ,, −−υυυφθθ2ba −1 + ∂∂∂∂t υυυθθφ st= (7) 3 ∂∂φφφθFF()ttt,, φθ () ttt ,, +=avtI3υυθθ3 ,,,, υυ θφ ∈∈ ∂∂υφ v where φφυυφθFF()ttt123,,= φθ() φ ,; t 1 θ ,;, tvt 2 3 is the joint characteristic function of the random variable (vector) (φθ(ttFt12),,( ) W ( 3)). Moreover the additional condition holds

φφφθFF()()()0,ttvt12 ;0, ; ,==⋅−∈ 0, ttvt 12 ;0, ; ,Y F() v δi t , v (8) and the initial condition

φ υ,0; υ ,0;0,t == φ υ ,0; υ ,0 υυ , , υυ , ∈ (9) φθφθF ()()()3 φθφθY φθφθ00 φθ

Hence we have derived a partial differential equation for the joint characteristic function

()φθ()ttFt,, ()W (). An alternative form of eq. (7)-(9) can be derived, by taking their Fourier transform. In this case we will have

∂ftts(),, φθF ∂∂⎡⎤⎡⎤3 +++++∈θθφφfφθFF()ttt,,() 2 b a13 a F f φθ () ttt ,, , t I (10) ∂∂t φθ⎣⎦ ∂⎣⎦⎢⎥ st=

where fφθFF()ttt123,,= f φθ (φθ ,;,;, t 1 t 2 Ft 3 ) is the joint probability density function of the random variable (vector) ()φθ()ttFt12,, ()W () 3. Moreover the additional condition holds

∞ iFν fttFtddfFtφθ,;,; , φθ== , νδ ⋅−∈i tdF ν , (11) ∫∫φθFFF()()()12 ∫ e Y () −∞ and the initial condition

−1 ⎡ ⎤ fFtdFfφθF ()()φθ, 0; , 0; ,== φθ φθ , 0; , 0F Y φ θ() υυ φ , θ , φθ , ∈ ∫ ⎣⎢ 00 ⎦⎥ (12)

Using this information, we can additionally determine the probability structure of the second derivative φ()t by means of a simple calculation using eq. (4b). Having this information we can calculate the probability of capsizing, thus, determining the ship stability and safety.

213 APPENDIX III

References

BELENKY L.V. AND SEVASIANOV N.B., 2003, Stability and Safety of Ships, Volume II : Risk of Capsizing, Elsevier. LLOYD, GERMANISCHER., 1992, Vorschriften und Richtlinien des Germanischen Lloyd. Volume 1: Schiffstechnik, Teil 1, Kapitel 1, Selbstverlag des Germanischen Lloyd, Hamburg. INTERNATIONAL MARITIME ORGANIZATION. 1995, Code on Intact Stability for All Types of Ships Covered by IMO Instruments. Resolution A.749(18). IMO, London. KREUZER, E. AND WENDT, M., 2000, ‘Ship capsizing analysis using advanced hydrodynamic modelling’, Philosphical Transactions of Royal Society London A 358, 1835–1851. SPYROU, K. J. AND THOMPSON, J. M. T. (eds.), 2000, ‘The nonlinear dynamics of ships’, Philosophical Transactions of Royal Society London, Series A (Theme Issue) 358. SPYROU, K. J. AND THOMPSON, J. M. T., 2000, ‘The nonlinear dynamics of ship motions: A field overview and some recent developments’, Philosophical Transactions of Royal Society London, Series A 358, 1735–1760.