<<

B.Sc./Grad. Dip.: Probability Models and Time MAS programmes: Stochastic Models and

4 The properties of AR(1) and MA processes

4.1 Autocovariances and for an AR(1) process

We derive the autocovariance and functions for an AR(1) process {Yt}, using two alternative methods. The first method is based upon the use of the model equation,

Yt = φYt−1 + ²t, t ∈ Z. (1)

Recall the infinite moving expression, X∞ i Yt = φ ²t−i, (2) i=0 derived in Section 3.7. The representation was shown to hold in the sense of square convergence for |φ| < 1. However, as we will see later, the representation also holds in the sense of “almost sure” or “probability one” convergence: with such an interpretation, it is apparent that Yt and ²t+i are uncorrelated for i ≥ 1. Therefore, it can be argued that Ys and ²t are uncorrelated for s < t, so that, since the process mean is zero,

E[Ys²t] = 0 s < t.

Squaring Equation (1) and taking expectations,

2 2 2 2 E[Yt ] = φ E[Yt−1] + 2φE[Yt−1²t] + E[²t ], i.e., 2 2 γ0 = φ γ0 + σ , 2 where σ = var(²t). Hence, σ2 γ = . (3) 0 1 − φ2

Multiplying Equation (1) by Yt−τ , where τ ≥ 1, and taking expectations, yields

γτ = φγτ−1 τ ≥ 1. (4)

Equation (4) together with the initial condition of Equation (3) has the solution

σ2 γ = φτ τ ≥ 0. (5) τ 1 − φ2

Recalling that ρτ = γτ /γ0, we obtain

τ ρτ = φ τ ≥ 0. (6)

1 Note the geometric decline of the autocorrelation function. If φ < 0 then the autocorrelation function oscillates and has negative correlation at lag 1. A slight variant of this method for obtaining the expression for ρτ is to divide through in Equation (4) by γ0 to obtain the recurrence relation

ρτ = φρτ−1 τ ≥ 1. (7)

Equation (7) together with the initial condition ρ0 = 1 has the solution obtained previously as Equation (6). Using the symmetry property that γ−τ = γτ and ρ−τ = ρτ , we may, if we wish, extend the of the values of τ in the solutions for the autocovariance function {γτ } and the autocorrelation function {ρτ }. Thus, for example, we may write

|τ| ρτ = φ τ ∈ Z.

An alternative approach to finding the expressions for {γτ } and {ρτ } is based upon use of the infinite expression of Equation (2). For τ ≥ 0,

γ = E[Y Y ] τ " t t−τ # X∞ X∞ i j = E φ ²t−i φ ²t−τ−j i=0 j=0 " # X∞ X∞ i j−τ = E φ ²t−i φ ²t−j i=0 j=τ X∞ X∞ i+j−τ = φ E[²t−i²t−j] i=0 j=τ X∞ = φ2j−τ σ2 j=τ σ2 = φτ . 1 − φ2 Thus we have again the result of Equation (5). Note that in the above derivation we have used the properties of the autocovariance function of a white noise process as given in Equations (1) and (2) of Section 3.3.

4.2 The lag operator It is often convenient to use the lag operator L to characterize models and to carry out math- ematical manipulations. (An alternative terminology is backward shift operator with corre- sponding notation B.) The operator is defined by

LYt = Yt−1 t ∈ Z. Defining Lj to be the ‘j-fold’ composition of L, then we note that

j L Yt = Yt−j.

2 Equation (1) for the AR(1) model may be written as

(1 − φL)Yt = ²t.

The infinite moving average representation of Yt may be more simply derived using the formal- ism (and associated algebra) of the lag operator. −1 Yt = (1 − φL) ²t X∞ i i = φ L ²t i=0 X∞ i = φ ²t−i, i=0 assuming that the sum converges.

4.3 Linear processes Definition 4.3.1 (Linear Process) A process {Yt} that has the representation X∞ Yt = ψi²t−i, (8) i=0 where {²t} is a white noise process and {ψi} is a sequence of coefficients such that P∞ j=0 |ψj| < ∞ is referred to as a linear process. Proposition 4.3.2 (Convergence and Stationarity of Linear Process) P∞ For a linear process {Yt}, with representation Yt = i=0 ψi²t−i:

P∞ (a) Yt = i=0 ψi²t−i is well defined, in the sense that the right hand side is almost surely bounded (i.e. bounded with probability one);

Pn (b) the sequence of partial sums, i=0 ψi²t−i, converges to Yt in mean square as n −→ ∞;

(c) {Yt} is (weakly) stationary, with mean zero and autocovariance function given by γ = E[Y Y ] τ " t t−τ # X∞ X∞ = E ψi²t−i ψj²t−τ−j i=0 j=0 " # X∞ X∞ = E ψi²t−i ψj−τ ²t−j i=0 j=τ X∞ X∞ = ψiψj−τ E[²t−i²t−j] i=0 j=τ X∞ 2 = σ ψjψj−τ τ ≥ 0. (9) j=τ

3 In particular, X∞ 2 2 var(Yt) = γ0 = σ ψj . j=0 P∞ P∞ 2 But since j=0 |ψj| < ∞, then j=0 ψj < ∞, and so var(Yt) is finite. Finiteness of γτ for τ 6= 0 follows from the calculation used in Examples 3 Qu. 3. 2 Remarks 4.3.3 (i) A linear process {Yt}, as defined above, is sometimes said to be causal or a causal function of {²t}.

(ii) We see from the infinite moving average representation of Equation (8) that the value Ys of the process at time s depends only on {²t : t ≤ s} and not on {²t : t > s}.

(iii) As for the special case of the AR(1) process, Ys and ²t are uncorrelated for s < t, so that E[Ys²t] = 0, s < t, (10) a fact that we shall make use of later.

i (iv) In the special case of {ψi = φ : i ∈ N} we have the infinite moving average represen- P∞ tation (2) of the AR(1) process, and thus j=0 |ψj| < ∞ is satisfied if and only if |φ| < 1. It P∞ i is readily checked by substitution that the process {Yt} defined by Yt = i=0 φ ²t−i satisfies Equation (1). It follows that a viable (i.e. stationary) AR(1) process with autoregressive pa- rameter φ exists if |φ| < 1.

P∞ 2 (v) The square summability of the coefficients i.e. j=0 ψj < ∞ is actually both necessary and sufficient for mean square convergence of the linear process representation.

4.4 The moving average model Consider a market that every working day receives fresh information which affects the price of a certain commodity. Let Yt denote the price change on day t. The immediate effect of the information received on day t upon Yt is represented by ²t, where {²t} is assumed to be a white noise process. But there is also a residual effect, such that Yt is affected by the information received on the q previous days. A simple model represents {Yt} as a moving average process. A moving average process of order q, an MA(q) process, with zero mean is a process {Yt} which satisfies the relation Xq Yt = ²t + θi²t−i t ∈ Z, (11) i=1 2 where {²t} is a white noise process with mean zero and σ and where θq 6= 0. For convenience, we may define θ0 = 1 and rewrite Equation (11) as Xq Yt = θi²t−i t ∈ Z. (12) i=0

4 • Note that we are dealing here with a finite moving average, having a finite number of moving average parameters, as against the infinite moving average representation that we described in the discussion of the AR(1) process.

• Historically, “moving ”P were introduced rather differently, with the coefficients θi defined in such a way that θi = 1. Each Yt value is then a weighted average of the ²t values. The average “moves” as t moves through successive values.

An MA(q) process is a special case of a linear process as defined in Section 4.3, with   θi for 0 ≤ i ≤ q ψ = . i  0 for i > q

Because all but a finite number of the coefficients ψi are zero, an MA(q) process necessarily satisfies the summability conditions on the coefficients in the definition of a linear process. Thus, for all moving average parameter values, θ1, θ2, . . . , θq, Equation (11) or (12) defines a with mean zero. Applying the result of Equation (9) to the present case, for τ ≥ 0,

Xq 2 γτ = σ θjθj−τ . j=τ

Thus  2 Pq−τ  σ i=0 θiθi+τ for 0 ≤ τ ≤ q γ = . τ  0 for τ > q In particular, Xq 2 2 var(Yt) = γ0 = σ θi . i=0 The autocorrelation function is given by  Pq−τ Pq 2  i=0 θiθi+τ / i=0 θi for 0 ≤ τ ≤ q ρ = . τ  0 for τ > q

Note that, for an MA(q) process, there is a cut-off point in the autocorrelation function – all autocorrelations beyond the q-th are zero. By comparison, for the AR(1) process and, as we shall see, for more general autoregressive processes, all autocorrelations are generally non-zero but die away geometrically as a function of the lag. This fact may be borne in mind when we are examining the sample autocorrelation function of an observed time series and considering what model to fit to the .

5 4.5 The first order moving average process

In the special case of the MA(1) process {Yt}, which satisfies the equation

Yt = ²t + θ²t−1 t ∈ Z, (13) the autocorrelation function is given by

ρ0 = 1

θ ρ = 1 1 + θ2

ρτ = 0, τ ≥ 2. Note that if θ > 0 then the MA(1) process is smoother than a white noise process but that if θ < 0 then the MA(1) process is more jagged than a white noise process.

Using Equation (13) recursively,

Yt = ²t + θ(Yt−1 − θ²t−2) 2 = ²t + θYt−1 − θ (Yt−2 − θ²t−3) = ... Xn k n+1 = ²t − (−θ) Yt−k − (−θ) ²t−n−1. (14) k=1 Here we would like to take the limit as n → ∞ in some appropriate sense for Equation (14) to obtain the infinite order autoregressive representation of {Yt}, X∞ k Yt = − (−θ) Yt−k + ²t. (15) k=1 Let us consider again the concept of convergence in mean square.   Ã !2 Xn £ ¤  k  2(n+1) 2 E − (−θ) Yt−k + ²t − Yt = E θ ²t−n−1 k=1 = θ2(n+1)σ2, which tends to zero as n → ∞ if and only if |θ| < 1. As a matter of fact, it can be shown that equation (15) is an exact representation (with probability one), in which case (as we will see later) the MA(1) process is an invertible process: this is the case if and only if |θ| < 1.

Using the lag operator, we may rewrite Equation (13) as

Yt = (1 + θL)²t.

6 Formally inverting this relationship, we may write X∞ X∞ −1 k k k ²t = (1 + θL) Yt = (−θ) L Yt = Yt + (−θ) Yt−k, k=0 k=1 which is equivalent to Equation (15). As shown above, this procedure is justifiable if and only if |θ| < 1.

Note that the MA(1) processes with parameters θ and θ−1, respectively, have the same autocorrelation function, since θ−1 θ = . 1 + θ−2 1 + θ2 Hence, given the autocorrelation function of an MA(1) process, it is impossible to identify uniquely the parameter value. But if we impose the condition that the process must be in- vertible then there is a unique parameter value corresponding to the given autocorrelation function. The MA(1) process with |θ| = 1 is exceptional in that it is not invertible but may be uniquely identified from the autocorrelation function.

4.6 Invertibility Using the lag operator, we may write the general MA(q) model of Equation (11) in the form

Yt = θ(L)²t, (16) where θ(z) is the MA characteristic polynomial, defined by

2 q θ(z) = 1 + θ1z + θ2z + ... + θqz . (17) Thus θ(z) is the generating function of the moving average coefficients. The corresponding MA characteristic equation is 2 q 1 + θ1z + θ2z + ... + θqz = 0. (18) Definition 4.6.1 (Invertibility) A stationary process {Yt} is said to be invertible if it can be expressed in the form X∞ Yt = πkYt−k + ²t, (19) k=1 P∞ such that j=1 |πj| < ∞. For an MA(q) process, the following result can be established.

Theorem 4.6.2 (Invertibility condition for MA(q)) A necessary and sufficient condition for the moving average parameters θ1, θ2, . . . , θq of the MA(q) model to specify an invertible moving average process is that all the roots of the characteristic equation, Equation (18), lie strictly outside the unit circle in the complex plane (i.e., all the roots are greater than one in modulus/absolute value). 2

7 We see from the expression (19) that the value at time t of an invertible process can be expressed as a linear combination of the values of the process at previous time points, with the addition of the white noise term at time t. This makes it plausible that it should be possible to forecast future values of such a process in a straightforward manner, given knowledge of its history. Generally, for purposes of estimation and , the invertibility condition is imposed on stationary processes. Given the autocorrelation function of a MA process, there is generally a multiplicity of parameter values that will yield the given autocorrelations, but there is only one corresponding set of parameters which specifies an invertible MA process.

Assuming that a given MA(q) process is invertible, we may rewrite (19) as

²t = π(L)Yt, (20) where X∞ k π(z) = 1 − πkz . k=1 We may invert (16) to obtain −1 ²t = θ(L) Yt. (21) It follows from equations (20) and (21) that we may write

π(z) = θ(z)−1 or, equivalently, θ(z)π(z) = 1. In the special case of the MA(1) process, θ(z) = 1 + θz, π(z) = (1 + θz)−1, and the invertibility condition reduces to |θ| < 1.

8