<<

- Introduction

• Stochastic processes are processes that proceed randomly in time.

• Rather than consider fixed random variables X, Y, etc. or even

sequences of i.i.d random variables, we consider sequences X0, X1, X2, …. Where Xt represent some random quantity at time t.

• In general, the value Xt might depend on the quantity Xt-1 at time t-1, or even the value Xs for other times s < t. • Example: simple random walk .

week 2 1 Stochastic Process - Definition

• A stochastic process is a family of time indexed random variables Xt

where t belongs to an index set. Formal notation, { t : ∈ ItX } where I is an index set that is a subset of R. • Examples of index sets:

1) I = (-∞, ∞) or I = [0, ∞]. In this case Xt is a continuous time stochastic process.

2) I = {0, ±1, ±2, ….} or I = {0, 1, 2, …}. In this case Xt is a discrete time stochastic process.

• We use uppercase letter {Xt } to describe the process. A , {xt } is a realization or sample function from a certain process. • We use information from a time series to estimate parameters and

properties of process {Xt }. week 2 2 of a Process

• For any stochastic process with index set I, its probability distribution function is uniquely determined by its finite dimensional distributions.

•The k dimensional distribution function of a process is defined by

,...,XX ( 1 ,..., k )= ( t ≤ 1 ,..., ≤ xXxXPxxF kt ) 1 tt k 1 k

for any 1 ,..., k ∈ Itt and any real numbers x1, …, xk . • The distribution function tells us everything we need to know about

the process {Xt }.

week 2 3 Moments of Stochastic Process

• We can describe a stochastic process via its moments, i.e., 2 ()t ( t ),, ()⋅ XXEXEXE st etc. We often use the first two moments.

• The function of the process is (XE )= μtt .

2 • The function of the process is ()XVar = σ tt .

• The function between Xt , Xs is

( ,Cov st )= (( − μ )(XXEXX − μ sstt ) )

• The correlation function between Xt , Xs is

( ,Cov XX st ) ρ(), XX st = 22 σσst • These moments are often function of time. week 2 4 Stationary Processes

( ,..., XX ) • A process is said to be strictly stationary if 1 tt k has the same joint distribution as ( ,..., XX ) . That is, if t1 Δ+ tk Δ+

,...,XX ( 1 ,..., )= FxxF ,...,XXk ( 1 ,..., xx k ) 1 tt k t1 Δ+ tk Δ+

2 •If {Xt } is a strictly and ( XE t ) ∞< then, the mean function is a constant and the variance function is also a constant.

• Moreover, for a strictly stationary process with first two moments finite, the covariance function, and the correlation function depend only on the time difference s.

• A trivial example of a strictly stationary process is a sequence of i.i.d random variables.

week 2 5 Weak Stationarity

• Strict stationarity is too strong of a condition in practice. It is often

difficult assumption to assess based on an observed time series x1,…,xk. • In time series analysis we often use a weaker sense of stationarity in terms of the moments of the process. • A process is said to be nth-order weakly stationary if all its joint moments up to order n exists and are time invariant, i.e., independent of time origin. • For example, a second-order weakly stationary process will have constant mean and variance, with the covariance and the correlation being functions of the time difference along. • A strictly stationary process with the first two moments finite is also a second-ordered weakly stationary. But a strictly stationary process may not have finite moments and therefore may not be weakly stationary.

week 2 6 The and Functions

• For a stationary process {Xt }, with constant mean μ and constant variance σ2. The covariance between Xt and Xt+s is

γ ()= ( ,cov +stt )= (( t − μ)(XXEXXs s − μ))

• The correlation between Xt and Xt+s is ( ,cov XX ) γ (s) ρ()s = st = var()t var()XX s γ ()0

Where var()t = var(XX +st )= γ ( ).0 • As functions of s, γ(s) is called the autocovariance function and ρ(s) is called the autocorrelation function (ATF). They represent the

covariance and correlation between Xt and Xt+s from the same process, separated only by s time lags.

week 2 7 Properties of γ(s) and ρ(s)

• For a stationary process, the autocovariance function γ(s) and the autocorrelation function ρ(s) have the following properties:

¾ γ ()= X t (ρ )( )= .10;var0 ¾ ρ()s ≤≤− 11 .

¾ γ ()γ −= (ss ) ρ(sand )= ρ(− s) . ¾ The autocovariance function γ(s) and the autocorrelation function ρ(s) are positive semidefinite in the sense that n n n n () ∑∑ ji γααji ≥− and0 ∑∑ ji ραα()ji ≥− 0 i==11j i==11j

for any real numbers α α 21 α n .,...,,

week 2 8

• A correlogram is a plot of the autocorrelation function ρ(s) versus the lag s where s = 0,1, ….

•Example…

week 2 9 Partial Autocorrelation Function

• Often we want to investigate the dependency / association between

Xt and Xt+k adjusting for their dependency on Xt+1, Xt+2,…, Xt+k-1.

• The conditional correlation Corr(Xt , Xt+k | Xt+1, Xt+2,…, Xt+k-1) is usually referred to as the in time series analysis.

• Partial autocorrelation is usually useful for identifying autoregressive models.

week 2 10

• A stochastic process is said to be a normal or Gaussian process if its joint probability distribution is normal.

• A Gaussian process is strictly and weakly stationary because the is uniquely characterized by its first two moments.

• The processes we will discuss are assumed to be Gaussian unless mentioned otherwise.

• Like other areas in , most time series results are established for Gaussian processes.

week 2 11 White Noise Processes

• A process {Xt} is called white noise process if it is a sequence of uncorrelated random variables from a fixed distribution with constant mean μ (usually assume to be 0) and constant variance σ2.

• A white noise process is stationary with autocovariance and autocorrelation functions given by ….

• A white noise process is Gaussian if its joint distribution is normal.

week 2 12 Estimation of the mean

• Given a single realization {xt} of a stationary process {Xt}, a natural

estimator of the mean ( XE t ) = μ is the sample mean 1 n x = ∑ xt n t=1 which is the time average of n observations.

• It can be shown that the sample mean is unbiased and consistent estimator for μ.

week 2 13 Sample Autocovariance Function

• Given a single realization {xt} of a stationary process {Xt}, the sample autocovariance function given by 1 =kn ˆ γ ()k ∑()t ()+kt −−= xxxx n t=1

is an estimate of the autocivariance function.

week 2 14 Sample Autocorrelation Function

• For a given time series {xt}, the sample autocorrelation function is given by =kn ()()−− xxxx ∑ t +kt γˆ()k ˆ t=1 ρ()k = n = . 2 γˆ()0 ∑()t − xx t=1 • The sample autocorrelation function is non-negative definite.

• The sample autocovariance and autocorrelation functions have the same properties as the autocovariance and autocorrelation function of the entire process.

week 2 15 Example

week 2 16