Ch. 23 Cointegration 1 Introduction An important property of (1) variables is that there can be linear combinations of theses variables that are I(O). If this is so then these variables are said to be cointegrated. Suppose that we consider two variables Yt and Xt that are I(1) (For example, Yt= Yt-1+ St and Xt= Xi-1+nt.)Then, Yt and Xt are said to be cointegrated if there exists a B such that Yt - BXt is I(O). What this mean is that the regression equation Yt= BX+ut make sense because Yt and Xt do not drift too far apart from each other over time. Thus, there is na long- run equilibrium relationship between them. If and Xt are not cointegrated, that is, Yt-BXt= ut is also I(1), then Yt and Xt would drift apart from each other over time. In this case the relationship between Yt and Xt that we obtain by regressing Yt and Xt would be spurious Let us continue the cointegration with the spurious regression setup in which Xt and Yt are independent random walks, consider what happens if we take a nontrivial linear combination of Xt and y alt+axT=a1Yt-1+a2Xt-1+a1St+a27t where a and ay are not both zero. We can write this as Zt= zt-1+U where Zt=aiT+a2 Xt and ut= a1St +a2nt. Thus, Zt is again a random walk process, as ut is i i.d. with mean zero and finite variance given that t and nt each are i.i.d. with mean zero and finite variance. No matter what coefficients a and a2 we choose, the resulting linear combination is again a random walk, hence an integrated or unit root process Now consider what happens when Xt is a random walk as before, but Yt is instead generated according to Yt= BX+ut, with ut again iid. By itself, Yt an integrated process, because Y-Y-1=(Xt-X1-1)3+t-u-1
Ch. 23 Cointegration 1 Introduction An important property of I(1) variables is that there can be linear combinations of theses variables that are I(0). If this is so then these variables are said to be cointegrated. Suppose that we consider two variables Yt and Xt that are I(1). (For example, Yt = Yt−1 + ζt and Xt = Xt−1 + ηt .) Then, Yt and Xt are said to be cointegrated if there exists a β such that Yt − βXt is I(0). What this mean is that the regression equation Yt = βXt + ut make sense because Yt and Xt do not drift too far apart from each other over time. Thus, there is na long-run equilibrium relationship between them. If Yt and Xt are not cointegrated, that is, Yt − βXt = ut is also I(1), then Yt and Xt would drift apart from each other over time. In this case the relationship between Yt and Xt that we obtain by regressing Yt and Xt would be spurious. Let us continue the cointegration with the spurious regression setup in which Xt and Yt are independent random walks, consider what happens if we take a nontrivial linear combination of Xt and Yt : a1Yt + a2Xt = a1Yt−1 + a2Xt−1 + a1ζt + a2ηt , where a1 and a2 are not both zero. We can write this as Zt = Zt−1 + vt , where Zt = a1Yt + a2Xt and vt = a1ζt + a2ηt . Thus, Zt is again a random walk process, as vt is i.i.d. with mean zero and finite variance, given that ζt and ηt each are i.i.d. with mean zero and finite variance. No matter what coefficients a1 and a2 we choose, the resulting linear combination is again a random walk, hence an integrated or unit root process. Now consider what happens when Xt is a random walk as before, but Yt is instead generated according to Yt = βXt + ut , with ut again i.i.d.. By itself, Yt is an integrated process, because Yt − Yt−1 = (Xt − Xt−1)β + ut − ut−1, 1
so that t-1 -1+Et where Et= Bnt +ut-ut-I is readily verified to be I(O)process Despite the fact that both Xt and Yt are integrated processes, the situation is very different from that considered at last chapter. Here, there is indeed a linear combination of Xt and Yt that are not an integrated process: putting a1= l and a2=-6 we have alT+a,Xt=Yt-BXt t which is i i d. This is an example of a pair (Xt, Yi of cointegrated process The concept of cointegration was introduced by Granger(1981). This pa per and that of Engle and Granger(1987) have had a major impact on modern econometrics. Following Engle and Granger(1987), we have the definition of cointegration formally as follows Definition 1 The components of the vector xt are said to be co-integrated of order d, b, de- noted xt N cI(d,b), if (a). all components of xt are I(d) (b). there exists a vector a( 0)so that at =a'xtNI(d-6),b>0. The vector a is called the co-integrating vector For ease of exposition, only the value d= l and b= l will be considered in this chapter. For the case that d and b are fractional value, this is called fractional cointegration. We will consider this case in Chapter 25 Clearly, the cointegrating vector a is not unique, for if a'xt is I(0), then so is ba'xt for any nonzero scalar b; if a is a cointegrating vector, then so is ba. If xt has h components, then there be more than one cointegrating vector a Indeed, there may be h k linear independent(k x 1) vectors(a1, a2 an )such
so that Yt = Yt−1 + βηt + ut − ut−1 = Yt−1 + εt , where εt = βηt + ut − ut−1 is readily verified to be I(0) process. Despite the fact that both Xt and Yt are integrated processes, the situation is very different from that considered at last chapter. Here, there is indeed a linear combination of Xt and Yt that are not an integrated process: putting a1 = 1 and a2 = −β we have a1Yt + a2Xt = Yt − βXt = ut , which is i.i.d. This is an example of a pair {Xt , Yt} of cointegrated process. The concept of cointegration was introduced by Granger (1981). This paper and that of Engle and Granger (1987) have had a major impact on modern econometrics. Following Engle and Granger (1987), we have the definition of cointegration formally as follows. Definition 1: The components of the vector xt are said to be co-integrated of order d, b, denoted xt ∼ CI(d, b), if (a). all components of xt are I(d); (b). there exists a vector a(6= 0) so that zt = a 0xt ∼ I(d − b), b > 0. The vector a is called the co-integrating vector. For ease of exposition, only the value d = 1 and b = 1 will be considered in this chapter. For the case that d and b are fractional value, this is called fractional cointegration. We will consider this case in Chapter 25. Clearly, the cointegrating vector a is not unique, for if a 0xt is I(0), then so is ba 0xt for any nonzero scalar b; if a is a cointegrating vector, then so is ba. If xt has k components, then there may be more than one cointegrating vector a. Indeed, there may be h < k linear independent (k×1) vectors (a1, a2,..., ah) such 2
that a'xt is a I(O)(h x 1) vector, where A' is the following (h x k)matrix Again, the vector(a1, a2,, an)are not unique; if A'xt is a I(O), then for any nonzero(1 x h) vector b, the scalar b'A'xt is also I(0). Then the(k x 1) vector given by b'a could also be described as a cointegrating vector Suppose that there exists an(h x k) matrix a whose rows are linearly inde- pendent such that Axt is a(h 1)I(0)vector. Suppose further that if c'is any (1 x k) vector that is linearly independent of the rows of A, then c'xt is a I(1) scalar. Then we say that there are exactly h cointegrating relations among the elements of xt and that(a1, a,,, an) form a basis for the space of the cointe- grating vectors Ex Let P denote an index of the price level in the United States, P* a price index for Italy, and St the exchange rate between the currency. Then purchasing power parity holds that Pt= S,P or. taking logarithms Pt=Stt pt where pt≡ log Pt,st≡ log St,andp≡logP. In equilibrium we need p St-Pt=0. However, in practice, error in measuring price, transportation costs, and differences in quality prevent purchasing power parity from holding exactly at every date t. A weaker form of the hypothesis is that the variable zt defined 2t =pt Pt I(O), even though the individual elements yt=(pt St p)are all I(1). In this case, we have a single cointegrating vector a=(1-1-1). The term at =ayu
that A0xt is a I(0) (h × 1) vector, where A0 is the following (h × k) matrix: A0 = a 0 1 a 0 2 . . . a 0 h . Again, the vector (a1 , a2,..., ah) are not unique; if A0xt is a I(0), then for any nonzero (1 × h) vector b 0 , the scalar b 0A0xt is also I(0). Then the (k × 1) vector φ given by φ 0 = b 0A0 could also be described as a cointegrating vector. Suppose that there exists an (h × k) matrix A0 whose rows are linearly independent such that A0xt is a (h × 1) I(0) vector. Suppose further that if c 0 is any (1 × k) vector that is linearly independent of the rows of A0 , then c 0xt is a I(1) scalar. Then we say that there are exactly h cointegrating relations among the elements of xt and that (a1, a2,..., ah) form a basis for the space of the cointegrating vectors. Example: Let Pt denote an index of the price level in the United States, P ∗ t a price index for Italy, and St the exchange rate between the currency. Then purchasing power parity holds that Pt = StP ∗ t , or, taking logarithms, pt = st + p ∗ t , where pt ≡ log Pt , st ≡ log St , and p ∗ t ≡ log P ∗ t . In equilibrium we need pt − st − p ∗ t = 0. However, in practice, error in measuring price, transportation costs, and differences in quality prevent purchasing power parity from holding exactly at every date t. A weaker form of the hypothesis is that the variable zt defined by zt = pt − st − p ∗ t is I(0), even though the individual elements yt = (pt st p ∗ t ) 0 are all I(1). In this case, we have a single cointegrating vector a = (1 −1 −1)0 . The term zt = a 0yt 3
is interpreted as the equilibrium error; although it is not always zero, but it can not be apart from zero too often and too far to make sense the equilibrium concept
is interpreted as the equilibrium error; although it is not always zero, but it can not be apart from zero too often and too far to make sense the equilibrium concept. 4
2 Granger Representation Theorem Let each elements of the(k x 1) vector, yt is I(1)with the(k x h)cointegrating matrix, A, such that each elements of Ayt is I(0). Then Granger(1983) have the following fundamental results when y are cointegrated 2.1 Implication of Cointegration For the VMA Represen- tation We now discuss the general implications of cointegration for the moving average representation of a vector system. Since it is assumed the Ayt is I(0, let d= (△yt), and def Suppose that ut has the Wold representation ut=Et+业1et-1+业2Et-1+….=业(L)et, where E(Et)=0 and E(EET) Q for t=T 0 otherwise Let y(1)denotes the(k x k) matrix polynomial y(a)evaluated at z= 1; that y(1)≡Ik+业1+业2+业 Then the following holds (a)A亚(1)=0 (b).A6=0. To verify this claim, note that as long as sys so is absolutely summabl the difference equation(1) implies that(from multivariate B-n decomposition) yt=yo+6·t+u1+u2+…+ut yo+0:t+(1)·(e1+e2+…+Et)+nt-7
2 Granger Representation Theorem Let each elements of the (k × 1) vector, yt is I(1) with the (k × h) cointegrating matrix, A, such that each elements of A0yt is I(0). Then Granger (1983) have the following fundamental results when y are cointegrated. 2.1 Implication of Cointegration For the VMA Representation We now discuss the general implications of cointegration for the moving average representation of a vector system. Since it is assumed the 4yt is I(0), let δ ≡ E(4yt), and define ut = 4yt − δ. (1) Suppose that ut has the Wold representation ut = εt + Ψ1εt−1 + Ψ2εt−1 + ... = Ψ(L)εt , where E(εt) = 0 and E(εtε 0 τ ) = Ω for t = τ 0 otherwise. Let Ψ(1) denotes the (k × k) matrix polynomial Ψ(z) evaluated at z = 1; that is, Ψ(1) ≡ Ik + Ψ1 + Ψ2 + Ψ3 + ..., Then the following holds. (a). A0Ψ(1) = 0, (b). A0δ = 0. To verify this claim, note that as long as {sΨs} ∞ s=0 is absolutely summable, the difference equation (1) implies that (from multivariate B-N decomposition): yt = y0 + δ · t + u1 + u2 + ... + ut (2) = y0 + δ · t + Ψ(1) · (ε1 + ε2 + ... + εt) + ηt − η0 , (3) 5