Академический Документы
Профессиональный Документы
Культура Документы
First we test to see if variables are stationary I(0). If not they are assumed to have a unit root and be I(1). If a set of variables are all I(1) they should not be estimated using ordinary regression analysis, but between them there may be one or more equilibrium relationships. We can both estimate how many and what they are (called cointegrating vectors) using Johansens technique.
If a set of variables are found to have one or more cointegrating vectors then a suitable estimation technique is a VECM (Vector Error Correction Model) which adjusts to both short run changes in variables and deviations from equilibrium. In what follows we work back to front. Starting with the VECMs, then Johansens technique than stationarity.
We have data on monthly unemployment rates in Indiana, Illinois, Kentucky, and Missouri from January 1978 through December 2003. We suspect that factor mobility will keep the unemployment rates in equilibrium. The following graph plots the data.
Note the form of the above line to draw the line graph; then the variables which will be plotted; finally t the time variable against which they are all plotted
For further info press the help key, then line
2
1980m1
10
12
1985m1
2000m1
2005m1
The graph shows that although the series do appear to move together, the relationship is not that. There are periods when Indiana has the highest rate and others when Indiana has the lowest rate. Although the Kentucky rate moves closely with the other series for most of the sample, there is a period in the mid-1980s when the unemployment rate in Kentucky does not fall at the same rate as the other series. We will model the series with two cointegrating equations and no linear or quadratic time trends in the original series. For now we use the noetable option to suppress displaying the short-run estimation table.
Cointegrating equations Equation _ce1 _ce2 Parms 2 2 chi2 133.3885 195.6324 P>chi2 0.0000 0.0000
Johansen normalization restrictions imposed beta _ce1 missouri indiana kentucky illinois _cons _ce2 missouri indiana kentucky illinois _cons -1.11e-16 1 .2059473 -1.51962 2.92857 . . .2718678 .2804792 .6743122 . . 0.76 -5.42 4.34 . . 0.449 0.000 0.000 . . -.3269038 -2.069349 1.606942 . . .7387985 -.9698907 4.250197 1 (omitted) .3493902 -1.135152 -.3880707 . .2005537 .2069063 .4974323 . 1.74 -5.49 -0.78 . 0.081 0.000 0.435 . -.0436879 -1.540681 -1.36302 . .7424683 -.7296235 .5868787 Coef. Std. Err. z P>|z| [95% Conf. Interval]
Except for the coefficients on kentucky in the two cointegrating equations and the constant term in the first, all the parameters are significant at the 5% level. We can refit the model with the Johansen normalization and the overidentifying constraint that the coefficient on kentucky in the second cointegrating equation is zero.
constraint 1 [_ce1]missouri = 1 constraint 2 [_ce1]indiana = 0 constraint 3 [_ce2]missouri = 0 constraint 4 [_ce2]indiana = 1 constraint 5 [_ce2]kentucky = 0 vec missouri indiana kentucky illinois, trend(rconstant) rank(2) lags(4) noetable bconstraints(1/5)
constraint 1 [_ce1]missouri = 1
Constraint number 1, [_ce1] tells us which equation and missouri=1 sets constraint.
beta _ce1 missouri indiana kentucky illinois _cons _ce2 missouri indiana kentucky illinois _cons
. .0907071 .6448924
. -1.492048 1.67305
. -1.136483 4.200982
The test of the overidentifying restriction does not reject the null hypothesis that the restriction is valid, and the p-value on the coefficient on kentucky in the first cointegrating equation indicates that it is not significant. We will leave the variable in the model and attribute the lack of significance to whatever caused the kentucky series to temporarily rise above the others from 1985 until 1990, though we could instead consider removing kentucky from the model.
Next, we look at the estimates of the adjustment parameters. In the output below, we replay the previous results. vec missouri indiana kentucky illinois, trend(rconstant) rank(2) lags(4) bconstraints(1/5)
Interpretation
D_missouri _ce1 L1. _ce2 L1. -.0683152 .0405613 .0185763 .0112417 -3.68 3.61 0.000 0.000 -.1047242 .018528 -.0319063 .0625946
If the error term in the first cointegration relation is positive unemployment in Missouri FALLS.
If the error term in the second cointegrating regression is positive then unemployment in Missouri INCREASES.
-.0683152
.0185763
-3.68
0.000
-.1047242
-.0319063
As we can see from the regression this is what we get. D_missouri is the change in unemployment in Missouri i.e. DUMt = Umt Umt-1 The coefficient on _ce1 L1 (_ce1 : the error term from the first cointegrating regression; L1 lagged one period) is -0.068 and significant at the 1% level.
Thus if in period t-1 the error term in _ce1 was positive, which we can see can be seen as unemployment in Missouri being too high compared to the equilibrium relationship with the other two states, then it will fall. The bigger the (negative) coefficient on _ce1 L1 the more rapid is the correction. If it = -1 then the entire error is corrected for in the following period.
_ce2 missouri indiana kentucky illinois _cons (omitted) 1 (omitted) -1.314265 2.937016 . .0907071 .6448924 . -14.49 4.55 . 0.000 0.000 . -1.492048 1.67305 . -1.136483 4.200982
_ce2 L1.
.0405613
.0112417
3.61
0.000
.018528
.0625946
-.0342096 .0325804
.0220955 .0133713
-1.55 2.44
0.122 0.015
-.0775159 .0063732
.0090967 .0587877
-.0342096 .0325804
.0220955 .0133713
-1.55 2.44
0.122 0.015
-.0775159 .0063732
.0090967 .0587877
Now this does not make much sense if rgw error term is positive unemployment in Indiana needs to fall to restore equilibrium. Yet the coefficient on it is positive indicating the opposite.
Another View
vec missouri indiana kentucky illinois, trend(rconstant) rank(2) lags(4) bconstraints(1/5) matrix cerr=e(beta) display cerr[1,1] display cerr[1,3] display cerr[1,5] display cerr[1,9] drop cerr1 cerr2
matrix cerr=e(beta) saves the coefficients from the two cointgretaing regressions in a vector cerr. cerr[1,1] is the first, cerr[1,9] is the penultimate coefficient in the second equation Thus: display cerr[1,9] gives: -1.3142654, the coefficient on Illinois in _ce2
Now this: regress D.missouri LD.missouri LD.indiana LD.kentucky LD.illinois L2D.missouri L2D.indiana L2D.kentucky L2D.illinois L3D.missouri L3D.indiana L3D.kentucky L3D.illinois L.cerr1 L.cerr2 Is almost equivalent to this: vec missouri indiana kentucky illinois, trend(rconstant) rank(2) lags(4) bconstraints(1/5)
I say almost because the VEC estimates both equations jointly and the regressions are slightly different, but very slightly. Note to if we have a slightly different short run structure then the cointegrating vectors change which is a little unsatisfactory
For example compare: vec missouri indiana kentucky illinois, trend(rconstant) rank(2) lags(2) bconstraints(1/5) vec missouri indiana kentucky illinois, trend(rconstant) rank(2) lags(4) bconstraints(1/5)
vec missouri indiana kentucky illinois arkansas ten, trend(rconstant) rank(3) lags(3) but Tennessee still remains unrelated to anything
We can see from the map that Tennessee is on the South east fringe of this group and it would be interesting to bring in North Carolina, Alabama and Georgia.
Johansens methoodology
vecrank implements three types of methods for determining r, the number of cointegrating equations in a VECM. The first is Johansens trace statistic method. The second is his maximum eigenvalue statistic method. The third method chooses r to minimize an information criterion. All three methods are based on Johansens maximum likelihood (ML) estimator of the parameters of a cointegrating VECM.
webuse balance2 We have quarterly data on the natural logs of aggregate consumption, investment, and GDP inthe United States from the first quarter of 1959 through the fourth quarter of 1982. As discussed in King et al. (1991), the balanced-growth hypothesis in economics implies that we would expect to find two cointegrating equations among these three variables.
describe
storage variable name type gdp t inv consump y i c float int float float double double double display format %9.0g %tq %9.0g %9.0g %10.0g %10.0g %10.0g value label variable label
In this example, because the trace statistic at r = 0 of 46.1492 exceeds its critical value of 29.68, we reject the null hypothesis of no cointegrating equations. Similarly, because the trace statistic at r = 1 of 17.581 exceeds its critical value of 15.41, we reject the null hypothesis that there is one or fewer cointegrating equation. In contrast, because the trace statistic at r = 2 of 3.3465 is less than its critical value of 3.76, we cannot reject the null hypothesis that there are two or fewer cointegrating equations.
. vecrank y i c, lags(5) Johansen tests for cointegration Trend: constant Sample: 1960q2 - 1982q4 maximum rank 0 1 2 3 Number of obs = Lags = 5% trace critical statistic value 46.1492 29.68 17.5810 15.41 3.3465* 3.76 91 5
parms 39 44 47 48
Because Johansens method for estimating r is to accept as the actual r the first r for which the null hypothesis is not rejected, we accept r = 2 as our estimate of the number of cointegrating equations between these three variables. The * by the trace statistic at r = 2 indicates that this is the value of r selected by Johansens multiple-trace test procedure
parms 39 44 47 48
The output indicates that switching from the 5% to the 1% level changes the resulting estimate from r = 2 to r = 1.
1% critical value 35.65 20.04 6.65 1% critical value 25.52 18.63 6.65
The test statistics are often referred to as lambda trace and lambda max respectively
We print out both tests in this table the eigenvalue ones are in the second half of the table. The test is for r versus r+1 cointegrating vectors. In this example, because the trace statistic at r = 0 of 46.1492 exceeds its critical value of 29.68, we reject the null hypothesis of no cointegrating equations. Similarly, because the trace statistic at r = 1 of 17.581 exceeds its critical value of 15.41, we reject the null hypothesis that there is one or fewer cointegrating equation. In contrast, because the trace statistic at r = 2 of 3.3465 is less than its critical value of 3.76, we cannot reject the null hypothesis that there are two or fewer cointegrating equations. The net result is we conclude there are 2 cointegrating vectors.
Stationarity
Intuitively a variable is stationary (I(0) integrated to order nought) if its characteristics do not change over time, e.g. variance, covariance and mean is unchanging. Another way of looking at it is that <1 in the following equation for a variable y: Yt = Yt-1 We do not estimate the above, but subtract Yt-1 from both sides: Yt - Yt-1 = Yt-1 -Yt-1 =(-1)Yt-1 Now in this regression we test that (-1) is significantly negative, which implies <1. If we reject this we say y is I(1) and it has a unit root. In this case doing time series OLS, etc on a variable with variables that are I(1) results in bias. The Johansen method is a suitable alternative. The above is the Dickey Fuller test (DF), add lagged values of Yt - Yt-1 to get rid of serial correlation and we have the augmented Dickey Fuller test.
Stationarity
Among the earliest tests proposed is the one by Dickey and Fuller (1979), though most researchers now use an improved variant called the augmented DickeyFuller test instead of the original version. Other common unit-root tests implemented in Stata include the DFGLS test of Elliot Rothenberg, and Stock (1996) and the PhillipsPerron (1988) test.
webuse air2 dfuller air The test statistics is less negative than any of the critical values and hence we cannot reject the null hypothesis that the variable exhibits a unit root and is thus not stationary
Dickey-Fuller test for unit root Test Statistic Z(t) -1.748 Number of obs = 143
Interpolated Dickey-Fuller 1% Critical 5% Critical 10% Critical Value Value Value -3.496 -2.887 -2.577
Augmented Dickey-Fuller test for unit root Test Statistic Z(t) -6.936 1% Critical Value -4.027
Number of obs
140
. dfuller air, lags(3) Augmented Dickey-Fuller test for unit root Test Statistic Z(t) -1.536 1% Critical Value -3.497 Number of obs = 140
MacKinnon approximate p-value for Z(t) = 0.0000 D.air air L1. LD. L2D. L3D. _trend _cons Coef. -.5217089 .5572871 .095912 .14511 1.407534 44.49164 Std. Err. .0752195 .0799894 .0876692 .0879922 .2098378 7.78335 t -6.94 6.97 1.09 1.65 6.71 5.72 P>|t| 0.000 0.000 0.276 0.101 0.000 0.000 [95% Conf. Interval] -.67048 .399082 -.0774825 -.0289232 .9925118 29.09753 -.3729379 .7154923 .2693065 .3191433 1.822557 59.88575
The regression basically regresses the change in the variable (D.air) on lagged changes an the lagged value of air plus a constant and time trend. The inclusion of lagged D.Air values makes this the augmented Dickey-Fuller test, i.e. it is what differentiates ir from the Dickey Fuller test.
pperron air
Phillips and Perrons test statistics can be viewed as Dickey Fuller statistics that have been made robust to serial correlation by using the NeweyWest (1987) heteroskedasticity- and autocorrelation-consistent covariance matrix estimator.
143 4
Interpolated Dickey-Fuller 5% Critical 10% Critical Value Value -13.786 -2.887 -11.057 -2.577
Z(rho) is the main statistic we are interested in as it is similar to the ADF test statistic.
DFGLS Test
webuse lutkepohl2 dfgls dln_inv dfgls tests for a unit root in a time series. It performs the modified DickeyFuller t test (known as the DF-GLS test) proposed by Elliott, Rothenberg, and Stock (1996). Essentially, the test is an augmented DickeyFuller test, similar to the test performed by Statas dfuller command, except that the time series is transformed via a generalized least squares (GLS) regression before performing the test. Elliott, Rothenberg, and Stock and later studies have shown that this test has significantly greater power than the previous versions of the augmented DickeyFuller test.