
Introduction to Time Series Analysis. Lecture 5. Peter Bartlett www.stat.berkeley.edu/∼bartlett/courses/153-fall2010 Last lecture: 1. ACF, sample ACF 2. Properties of the sample ACF 3. Convergence in mean square 1 Introduction to Time Series Analysis. Lecture 5. Peter Bartlett www.stat.berkeley.edu/∼bartlett/courses/153-fall2010 1. AR(1) as a linear process 2. Causality 3. Invertibility 4. AR(p) models 5. ARMA(p,q) models 2 AR(1) as a linear process Let {Xt} be the stationary solution to Xt − φXt−1 = Wt, where 2 Wt ∼ W N(0, σ ). If |φ| < 1, ∞ j X = φ W − t X t j j=0 is the unique solution: • This infinite sum converges in mean square, since |φ| < 1 implies |φj| < ∞. P • It satisfies the AR(1) recurrence. 3 AR(1) in terms of the back-shift operator We can write Xt − φXt−1 = Wt ⇔ (1 − φB) Xt = Wt | φ{z(B) } ⇔ φ(B)Xt = Wt Recall that B is the back-shift operator: BXt = Xt−1. 4 AR(1) in terms of the back-shift operator Also, we can write ∞ j X = φ W − t X t j j=0 ∞ ⇔ X = φjBj W t X t j=0 | π{z(B) } ⇔ Xt = π(B)Wt 5 AR(1) in terms of the back-shift operator With these definitions: ∞ π(B)= φjBj and φ(B) = 1 − φB, X j=0 we can check that π(B)= φ(B)−1: ∞ ∞ ∞ π(B)φ(B)= φjBj(1 − φB)= φjBj − φjBj = 1. X X X j=0 j=0 j=1 Thus, φ(B)Xt = Wt ⇒ π(B)φ(B)Xt = π(B)Wt ⇔ Xt = π(B)Wt. 6 AR(1) in terms of the back-shift operator Notice that manipulating operators like φ(B), π(B) is like manipulating polynomials: 1 =1+ φz + φ2z2 + φ3z3 + ··· , 1 − φz provided |φ| < 1 and |z|≤ 1. 7 Introduction to Time Series Analysis. Lecture 5. 1. AR(1) as a linear process 2. Causality 3. Invertibility 4. AR(p) models 5. ARMA(p,q) models 8 AR(1) and Causality Let Xt be the stationary solution to Xt − φXt−1 = Wt, 2 where Wt ∼ W N(0, σ ). If |φ| < 1, ∞ j X = φ W − . t X t j j=0 φ = 1? φ = −1? |φ| > 1? 9 AR(1) and Causality If |φ| > 1, π(B)Wt does not converge. But we can rearrange Xt = φXt−1 + Wt 1 1 as X − = X − W , t 1 φ t φ t and we can check that the unique stationary solution is ∞ − X = − φ jW . t X t+j j=1 But... Xt depends on future values of Wt. 10 Causality A linear process {Xt} is causal (strictly, a causal function of {Wt}) if there is a 2 ψ(B)= ψ0 + ψ1B + ψ2B + ··· ∞ with |ψ | < ∞ X j j=0 and Xt = ψ(B)Wt. 11 AR(1) and Causality • Causality is a property of {Xt} and {Wt}. • Consider the AR(1) process defined by φ(B)Xt = Wt (with φ(B) = 1 − φB): φ(B)Xt = Wt is causal iff |φ| < 1 iff the root z1 of the polynomial φ(z) = 1 − φz satisfies |z1| > 1. 12 AR(1) and Causality • Consider the AR(1) process φ(B)Xt = Wt (with φ(B) = 1 − φB): If |φ| > 1, we can define an equivalent causal model, −1 Xt − φ Xt−1 = W˜ t, where W˜ t is a new white noise sequence. 13 AR(1) and Causality • Is an MA(1) process causal? 14 Introduction to Time Series Analysis. Lecture 5. 1. AR(1) as a linear process 2. Causality 3. Invertibility 4. AR(p) models 5. ARMA(p,q) models 15 MA(1) and Invertibility Define Xt = Wt + θWt−1 = (1+ θB)Wt. If |θ| < 1, we can write −1 (1 + θB) Xt = Wt 2 2 3 3 ⇔ (1 − θB + θ B − θ B + ··· )Xt = Wt ∞ j ⇔ (−θ) X − = W . X t j t j=0 That is, we can write Wt as a causal function of Xt. We say that this MA(1) is invertible. 16 MA(1) and Invertibility Xt = Wt + θWt−1 ∞ j If |θ| > 1, the sum (−θ) Xt−j diverges, but we can write Pj=0 −1 −1 Wt−1 = −θ Wt + θ Xt. Just like the noncausal AR(1), we can show that ∞ − W = − (−θ) jX . t X t+j j=1 That is, we can write Wt as a linear function of Xt, but it is not causal. We say that this MA(1) is not invertible. 17 Invertibility A linear process {Xt} is invertible (strictly, an invertible function of {Wt}) if there is a 2 π(B)= π0 + π1B + π2B + ··· ∞ with |π | < ∞ X j j=0 and Wt = π(B)Xt. 18 MA(1) and Invertibility • Invertibility is a property of {Xt} and {Wt}. • Consider the MA(1) process defined by Xt = θ(B)Wt (with θ(B)=1+ θB): Xt = θ(B)Wt is invertible iff |θ| < 1 iff the root z1 of the polynomial θ(z)=1+ θz satisfies |z1| > 1. 19 MA(1) and Invertibility • Consider the MA(1) process Xt = θ(B)Wt (with θ(B)=1+ θB): If |θ| > 1, we can define an equivalent invertible model in terms of a new white noise sequence. • Is an AR(1) process invertible? 20 Introduction to Time Series Analysis. Lecture 5. 1. AR(1) as a linear process 2. Causality 3. Invertibility 4. AR(p) models 5. ARMA(p,q) models 21 AR(p): Autoregressive models of order p An AR(p) process {Xt} is a stationary process that satisfies Xt − φ1Xt−1 −···− φpXt−p = Wt, 2 where {Wt}∼ W N(0, σ ). Equivalently, φ(B)Xt = Wt, p where φ(B) = 1 − φ1B −···− φpB . 22 AR(p): Constraints on φ Recall: For p = 1 (AR(1)), φ(B) = 1 − φ1B. This is an AR(1) model only if there is a stationary solution to φ(B)Xt = Wt, which is equivalent to |φ1|= 6 1. This is equivalent to the following condition on φ(z) = 1 − φ1z: ∀z ∈ R, φ(z) = 0 ⇒ z 6= ±1 equivalently, ∀z ∈ C, φ(z) = 0 ⇒ |z| 6= 1, where C is the set of complex numbers. 23 AR(p): Constraints on φ Stationarity: ∀z ∈ C, φ(z) = 0 ⇒ |z|= 6 1, where C is the set of complex numbers. φ(z) = 1 − φ1z has one root at z1 = 1/φ1 ∈ R. But the roots of a degree p> 1 polynomial might be complex. For stationarity, we want the roots of φ(z) to avoid the unit circle, {z ∈ C : |z| = 1}. 24 AR(p): Stationarity and causality Theorem: A (unique) stationary solution to φ(B)Xt = Wt exists iff p φ(z) = 1 − φ1z −···− φpz = 0 ⇒ |z| 6= 1. This AR(p) process is causal iff p φ(z) = 1 − φ1z −···− φpz = 0 ⇒ |z| > 1. 25 Recall: Causality A linear process {Xt} is causal (strictly, a causal function of {Wt}) if there is a 2 ψ(B)= ψ0 + ψ1B + ψ2B + ··· ∞ with |ψ | < ∞ X j j=0 and Xt = ψ(B)Wt. 26 AR(p): Roots outside the unit circle implies causal (Details) ∀z ∈ C, |z|≤ 1 ⇒ φ(z) 6= 0 ∞ 1 ⇔ ∃{ψ },δ> 0, ∀|z|≤ 1+ δ, = ψ zj. j φ(z) X j j=0 j j 1/j ⇒ ∀|z|≤ 1+ δ, |ψj z | → 0, |ψj | |z| → 0 ∞ 1 ⇒ ∃j , ∀j ≥ j , |ψ |1/j ≤ ⇒ |ψ | < ∞. 0 0 j 1+ δ/2 X j j=0 m So if |z|≤ 1 ⇒ φ(z) 6= 0, then S = ψ BjW converges in mean m X j t j=0 −1 square, so we have a stationary, causal time series Xt = φ (B)Wt. 27 Calculating ψ for an AR(p): matching coefficients 2 Example: Xt = ψ(B)Wt ⇔ (1 − 0.5B + 0.6B )Xt = Wt, so 1= ψ(B)(1 − 0.5B + 0.6B2) 2 2 ⇔ 1=(ψ0 + ψ1B + ψ2B + ··· )(1 − 0.5B + 0.6B ) ⇔ 1= ψ0, 0= ψ1 − 0.5ψ0, 0= ψ2 − 0.5ψ1 + 0.6ψ0, 0= ψ3 − 0.5ψ2 + 0.6ψ1, . 28 Calculating ψ for an AR(p): example ⇔ 1= ψ0, 0= ψj (j ≤ 0), 0= ψj − 0.5ψj−1 + 0.6ψj−2 ⇔ 1= ψ0, 0= ψj (j ≤ 0), 0= φ(B)ψj. We can solve these linear difference equations in several ways: • numerically, or • by guessing the form of a solution and using an inductive proof, or • by using the theory of linear difference equations. 29 Calculating ψ for an AR(p): general case φ(B)Xt = Wt, ⇔ Xt = ψ(B)Wt so 1= ψ(B)φ(B) p ⇔ 1=(ψ0 + ψ1B + ··· )(1 − φ1B −···− φpB ) ⇔ 1= ψ0, 0= ψ1 − φ1ψ0, 0= ψ2 − φ1ψ1 − φ2ψ0, . ⇔ 1= ψ0, 0= ψj (j < 0), 0= φ(B)ψj. 30 Introduction to Time Series Analysis. Lecture 5. 1. AR(1) as a linear process 2. Causality 3. Invertibility 4. AR(p) models 5. ARMA(p,q) models 31 ARMA(p,q): Autoregressive moving average models An ARMA(p,q) process {Xt} is a stationary process that satisfies Xt −φ1Xt−1 −···−φpXt−p = Wt +θ1Wt−1 +···+θqWt−q, 2 where {Wt}∼ W N(0, σ ). • AR(p) = ARMA(p,0): θ(B) = 1. • MA(q) = ARMA(0,q): φ(B) = 1. 32 ARMA processes Can accurately approximate many stationary processes: For any stationary process with autocovariance γ, and any k > 0, there is an ARMA process {Xt} for which γX (h)= γ(h), h = 0, 1,...,k. 33 ARMA(p,q): Autoregressive moving average models An ARMA(p,q) process {Xt} is a stationary process that satisfies Xt −φ1Xt−1 −···−φpXt−p = Wt +θ1Wt−1 +···+θqWt−q, 2 where {Wt}∼ W N(0, σ ).
Details
-
File Typepdf
-
Upload Time-
-
Content LanguagesEnglish
-
Upload UserAnonymous/Not logged-in
-
File Pages34 Page
-
File Size-