<<

ARMA Functions

• For a moving process, MA(q):

xt = wt + θ1wt−1 + θ2wt−2 + ··· + θqwt−q.

• So (with θ0 = 1)   γ(h) = cov xt+h, xt  q   q  X X = E  θjwt+h−j  θkwt−k j=0 k=0  q−h  2 X σ θ θ , 0 ≤ h ≤ q = w j j+h j=0  0 h > q.

1 • So the ACF is  q−h   X  θjθj+h  j=0  q , 0 ≤ h ≤ q ρ(h) = X 2  θj   j=0  0 h > q.

• Notes:

– In these expressions, θ0 = 1 for convenience.

– γ(q) 6= 0 but γ(h) = 0 for h > q. This characterizes MA(q).

2 • For an autoregressive process, AR(p):

xt = φ1xt−1 + φ2xt−2 + ··· + φpxt−p + wt.

• So   γ(h) = cov xt+h, xt  p   X = E  φjxt+h−j + wt+h xt j=1 p X   = φjγ(h − j) + cov wt+h, xt . j=1

3 • Because xt is causal, xt is wt+ a linear combination of wt−1, wt−2,... .

• So  2   σw h = 0 cov wt+h, xt = 0 h > 0.

• Hence p X γ(h) = φjγ(h − j), h > 0 j=1 and p X 2 γ(0) = φjγ(−j) + σw. j=1

4 2 • If we know the parameters φ1, φ2, . . . , φp and σw, these equa- tions for h = 0 and h = 1, 2, . . . , p form p + 1 linear equations in the p + 1 unknowns γ(0), γ(1), . . . , γ(p).

• The other autocovariances can then be found recursively from the equation for h > p.

• Alternatively, if we know (or have estimated) γ(0), γ(1), . . . , γ(p), they form p + 1 linear equations in the p + 1 parameters 2 φ1, φ2, . . . , φp and σw.

• These are the Yule-Walker equations.

5 • For the ARMA(p, q) model with p > 0 and q > 0:

xt =φ1xt−1 + φ2xt−2 + ··· + φpxt−p + wt + θ1wt−1 + θ2wt−2 + ··· + θqwt−q, a generalized set of Yule-Walker equations must be used.

• The models ARMA(0, q) = MA(q) are the only ones with a closed form expression for γ(h).

• For AR(p) and ARMA(p, q) with p > 0, the recursive equation that for h > max(p, q + 1), γ(h) is a sum of geomet- rically decaying terms, possibly damped oscillations.

6 • The recursive equation is p X γ(h) = φjγ(h − j), h > q. j=1

• What kinds of sequences satisfy an equation like this?

– Try γ(h) = z−h for some constant z.

– The equation becomes p  p  −h X −(h−j) −h X j −h 0 = z − φjz = z 1 − φjz  = z φ(z). j=1 j=1

7 • So if φ(z) = 0, then γ(h) = z−h satisfies the equation.

• Since φ(z) is a polynomial of degree p, there are p solutions, say z1, z2, . . . , zp.

• So a more general solution is p X −h γ(h) = clzl , l=1 for any constants c1, c2, . . . , cp.

• If z1, z2, . . . , zp are distinct, this is the most general solution; if some roots are repeated, the general form is a little more complicated.

8 • If all z1, z2, . . . , zp are real, this is a sum of geometrically decaying terms.

• If any root is complex, its complex conjugate must also be a root, and these two terms may be combined into geometri- cally decaying sine-cosine terms.

• The constants c1, c2, . . . , cp are determined by initial condi- tions; in the ARMA case, these are the Yule-Walker equa- tions.

• Note that the various rates of decay are the zeros of φ(z), the autoregressive operator, and do not depend on θ(z), the moving average operator.

9 • Example: ARMA(1, 1)

xt = φxt−1 + θwt−1 + wt.

• The recursion is

γ(h) = φγ(h − 1), h = 2, 3,...

• So γ(h) = cφh for h = 1, 2,... , but c 6= 1.

• Graphically, the ACF decays geometrically, but with a differ- ent value at h = 0.

10 ● 1.0 0.8

0.6 ●

● ● ●

0.4 ● ● ● ● ●

ARMAacf(ar = 0.9, ma −0.5, 24) ● ●

0.2 ● ● ● ● ● ● ● ● ● ● ● ●

5 10 15 20 25

Index

11 The Partial Autocorrelation Function

• An MA(q) can be identified from its ACF: non-zero to lag q, and zero afterwards.

• We need a similar tool for AR(p).

• The partial autocorrelation function (PACF) fills that role.

12 • Recall: for multivariate random variables X, Y, Z, the partial correlations of X and Y given Z are the correlations of:

– the residuals of X from its regression on Z; and

– the residuals of Y from its regression on Z.

• Here “regression” means conditional expectation, or best lin- ear prediction, based on population distributions, not a sam- ple calculation.

• In a time , the partial are defined as

φh,h = of xt+h and xt given xt+h−1, xt+h−2, . . . , xt+1.

13 • For an autoregressive process, AR(p):

xt = φ1xt−1 + φ2xt−2 + ··· + φpxt−p + wt,

• If h > p, the regression of xt+h on xt+h−1, xt+h−2, . . . , xt+1 is

φ1xt+h−1 + φ2xt+h−2 + ··· + φpxt+h−p

• So the residual is just wt+h, which is uncorrelated with xt+h−1, xt+h−2, . . . , xt+1 and xt.

14 • So the partial autocorrelation is zero for h > p:

φh,h = 0, h > p.

• We can also show that φp,p = φp, which is non-zero by as- sumption.

• So φp,p 6= 0 but φh,h = 0 for h > p. This characterizes AR(p).

15 The Inverse Autocorrelation Function

• SAS’s proc arima also shows the Inverse Autocorrelation Func- tion (IACF).

• The IACF of the ARMA(p, q) model

φ(B)xt = θ(B)wt is defined to be the ACF of the inverse (or dual) process (inverse) θ(B)xt = φ(B)wt.

• The IACF has the same property as the PACF: AR(p) is characterized by an IACF that is nonzero at lag p but zero for larger lags.

16 Summary: Identification of ARMA processes

• AR(p) is characterized by a PACF or IACF that is:

– nonzero at lag p;

– zero for lags larger than p.

• MA(q) is characterized by an ACF that is:

– nonzero at lag q;

– zero for lags larger than q.

• For anything else, try ARMA(p, q) with p > 0 and q > 0.

17 For p > 0 and q > 0:

AR(p) MA(q) ARMA(p, q) ACF Tails off Cuts off after lag q Tails off PACF Cuts off after lag p Tails off Tails off IACF Cuts off after lag p Tails off Tails off

• Note: these characteristics are used to guide the initial choice of a model; estimation and model-checking will often lead to a different model.

18 Other ARMA Identification Techniques

• SAS’s proc arima offers the MINIC option on the identify statement, which produces a table of SBC criteria for various values of p and q.

• The identify statement has two other options: ESACF and SCAN.

• Both produce tables in which the pattern of zero and non- zero values characterize p and q.

• See Section 3.4.10 in Brocklebank and Dickey. 19 options linesize = 80; ods html file = ’varve3.html’; varve; infile ’../data/varve.dat’; input varve; lv = log(varve); run; proc arima data = varve; title ’Use identify options to identify a good model’; identify var = lv(1) minic esacf scan; estimate q = 1 method = ml; estimate q = 2 method = ml; estimate p = 1 q = 1 method = ml; run;

• proc arima output