Академический Документы
Профессиональный Документы
Культура Документы
ECONOMETRICS 2
TERM PAPER
SAFIA ASLAM
2013-14
2|E C O N O M E T R I C S 2 T E R M P A P E R
CONTENTS
Chapter 8 Multiple Regression Analysis: The Problem of Inference R e s t r i c t e d Least Squares: Testing Linear Equality Restrictions W a l d Test T e s t i n g For Structural or Parameter Stability of Regression Models O The Chow Test C o m p u t e r Application using EViews Chapter 10 Multicollinearity: What Happens If The Regressors Are Correlated? The Nature of Multicollinearity Sources of Multicollinearity Practical Consequences of Multicollinearity Detection of Multicollinearity O High R2 but Few Significant T-Ratios O High Pair-wise Correlations among Regressors O Examination of Partial Correlations O Auxiliary Regressions O Tolerance and Variance Inflation Factor O Remedial Measures O Do Nothing Or O Follow Some Rules Of Thumb. O A Priori Information O Combining Cross-Sectional and Time Series Data O Dropping a Variable(S) and Specification Bias O Transformation of Variables O First Difference Form O Ratio Transformation
3|E C O N O M E T R I C S 2 T E R M P A P E R
A d d i t i on a l or New Data O Computer Application using EViews Chapter 12 Autocorrelation: What Happens If the Error Terms Are Correlated? The Nature of Autocorrelation Sources of Autocorrelation O Inertia. O Specification Bias: Excluded Variables Case O Specification Bias: Incorrect Functional Form O Cobweb Phenomenon O Lags. O Manipulation Of Data O Data Transformation Practical Consequences of Autocorrelation
4|E C O N O M E T R I C S 2 T E R M P A P E R
Detection of Autocorrelation G r a p h i c a l Method T h e Runs Test D u r b i n Watson D Test A General Test of Autocorrelation: The BreuschGodfrey (Bg) Test R e m e d i a l Measures of Autocorrelation N e w e y West Method. G e n e r a l i z e d Least-Square (G|LS) Method. Correcting For (Pure) Autocorrelation: The Method of Generalized Least Squares (GLS) O When Is Known: O When Is Not Known: The First-Difference Method: O Berenblutt Webb Test, Based On DurbinWatson D Statistic: Estimated From The Residuals: Theil-Nagar Estimate Based On D Statistic: Estimating : The CochraneOrcutt (CO) Iterative Procedure: Estimating : Durbins Two-Step Method: The Durbin h Statistic Computer Application using EViews Chapter 11 Heteroscedasticity: What Happens If the Error Variance Is Non constant? The Nature of Heteroscedasticity Sources of Heteroscedasticity Error-Learning Models Data Collecting Techniques Improve Outliers.
2 i
Is Likely To Decrease.
Other Sources of Heteroscedasticity: O Incorrect Data Transformation (E.G., Ratio or First Difference Transformations) O Incorrect Functional Form (E.G., Linear Versus LogLinear Models). Practical Consequences of Heteroscedasticity Detection of Heteroscedasticity Informal Methods
5|E C O N O M E T R I C S 2 T E R M P A P E R
Nature Of The Problem Graphical Method Formal Methods O Park Test O Glejser Test O Goldfeld-Quandt Test W h i t e s General Heteroscedasticity Test K o e n k e r Bassett (Kb) Test Remedial Measures 1. When i2 is known The method of weighted least square 2. When i2 is not known Whites heteroscedasticity-Consistent Variances and Standard Errors Plausible Assumptions about Heteroscedasticity Pattern.
Assumption 1: The Error Variance Is Proportional To Xi2 Assumption 2: The Error Variance Is Proportional To Xi. The Square Root i Transformation Assumption 3: The Error Variance Is Proportional To the Square Of The Mean Value Of Y. Assumption 4: A Log Transformation 3.
1. The Nature Of Simultaneous-Equation Models 2. The Identification Problem 3. Rules For Identification The Order Condition Of Identifiability The Rank Condition Of Identifiability Hausman Specification Test The Method Of Indirect Least Squares (ILS): A Just Identified Equation The Method Of Two-Stage Least Squares (2SLS): An Over identified
6|E C O N O M E T R I C S 2 T E R M P A P E R
Equation The Method Of Three-Stage Least Squares (3SLS) Using EViews The Granger Test Computer Application using EViews
The Nature of Dummy Variable Caution In the Use of Dummy Variables Analysis of Variance (ANOVA) Model Analysis of Covariance (ANCOVA) Models. Computer Application using EView
7|P a g e
WALD Test: Wald test is used to test the validity of the linear restriction imposed on the parameters. There are occasions where economic theory may suggest that the coefficients in a regression model satisfy some linear equality restrictions.
Y=KLe
Where Yi = output, Li = labor input, and Ki = capital input. Written in log form, the equation becomes
LnYi = 1 + 2 ln Ki + 3 ln Li + i (8.1)
Now if there are constant returns to scale (equi proportional change in output for an equiproportional change in the inputs), economic theory would suggest that
2 + 3 = 1
This is an example of a linear equality restriction. TESTING LINEAR RESTRICTIONS IN EVIEWS:
Step1: First Regress the Model (8.1) equation. Quick estimate equation write: log(Y) C log(K) log(L)
Dependent Variable: LOG(Y) Method: Least Squares Date: 08/21/13 Time: 13:41
8|E C O N O M E T R I C S 2 T E R M P A P E R
Sample: 1960 2012 Included observations: 53 Coefficien t 10.94926 3.734688 -1.271933 0.996592 0.996456 0.025746 0.033143 120.2927 0.054050
Variable C LOG(K) LOG(L) R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Prob. 0.0000 0.0000 0.0000 18.36495 0.432456 -4.426138 -4.314612 7310.653 0.000000
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
Step2: First Regress the Model (8.1) equation Regression window View Coefficient Tests Wald-Coefficients Restrictions: Write: C(2) + C(3) = 1
Test Statistic
Value
df
Probability
F-statistic Chi-square
319.2578 319.2578
(1, 50) 1
0.0000 0.0000
9|E C O N O M E T R I C S 2 T E R M P A P E R
Normalized Restriction (= 0)
Value
Std. Err.
-1 + C(2) + C(3)
1.462755
0.081865
STEP 1: First Regress the (8.1) equation. Quick estimate equation write: log(Y) C log(K) log(L) And Obtain RSSUR Dependent Variable: LOG(Y) Method: Least Squares Date: 08/21/13 Time: 13:41 Sample: 1960 2012 Included observations: 53 Coefficien t 10.94926 3.734688 -1.271933 0.996592 0.996456
10 | E C O N O M E T R I C S 2 T E R M P A P E R
STEP 2: First Regress the (8.2) equation. Quick estimate equation write: log(Y/L) C log(K/L) And Obtain RSSR
Dependent Variable: LOG(Y/L) Method: Least Squares Date: 08/21/13 Time: 13:48 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C LOG(K/L)
18.42422 5.936631
0.119150 0.170981
154.6300 34.72092
0.0000 0.0000
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic
11 | E C O N O M E T R I C S 2 T E R M P A P E R
Durbin-Watson stat
0.035469
Prob(F-statistic)
0.000000
STEP 3: Apply the F Test of RSS Version. F = { (RSSR RSSUR) /M} / {RSSUR / (N-K)}
And compare it with F critical Values at m and (n-k) degrees of freedom; if | FCal | > | FCritical | then Restriction is invalid and vice versa
TESTING FOR STRUCTURAL OR PARAMETER STABILITY OF REGRESSION MODELS: THE CHOW TEST When we use a regression model involving time series data, it may happen that there is a structural change in the relationship between the regressand Y and the regressors. By structural change, we mean that the values of the parameters of the model do not remain the same through the entire time period. USING E VIEWS:
Step 1: First Regress the Model (8.3) equation (with n = 34); Quick estimate equation write: Sav C Yd
12 | E C O N O M E T R I C S 2 T E R M P A P E R
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C YD
-15.73656 0.771388
1.331322 0.022667
-11.82025 34.03118
0.0000 0.0000
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
Step 2: First Regress the Model (8.3) equation. Regression window View Stability Tests Chow Breakpoint Test: Write Breakpoint Year in the Box: 1980
303.2648 137.4620
Probability Probability
0.000000 0.000000
13 | E C O N O M E T R I C S 2 T E R M P A P E R
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1 X2 X3 X4 X5
0.997288 0.997000
58.49076 5.387357
14 | E C O N O M E T R I C S 2 T E R M P A P E R
THE NATURE OF MULTICOLLINEARITY The term multicollinearity is due to Ragnar Frisch. Originally it meant the existence of a perfect, or exact, linear relationship among some or all explanatory variables of a regression model. X1 + X2 . + Xk = 0 X1 + X2 . + Xk + i = 0 (10.1.1) (10.1.2)
Why does the classical linear regression model assume that there is no multicollinearity among the Xs? The reasoning is this: If multicollinearity is perfect in the sense of (10.1.1), the regression coefficients of the X variables are indeterminate and their standard errors are infinite. If multicollinearity is less than perfect, as in (10.1.2), the regression coefficients, although determinate, possess large standard errors (in relation to the coefficients themselves), which means the coefficients cannot be estimated with great precision or accuracy.
15 | E C O N O M E T R I C S 2 T E R M P A P E R
SOURCES OF MULTICOLLINEARITY There are several sources of multicollinearity. As Montgomery and Peck note, multicollinearity may be due to the following factors; 1. The data collection method employed, for example, sampling over a limited range of the values 2.
taken by the regressors in the population Constraints on the model or in the population being sampled. For example, in the regression of electricity consumption on income (X2) and house size (X3) there is a physical constraint in the population in that families with higher incomes generally have larger homes than families with lower incomes. Model specification, for example, is adding polynomial terms to a regression model, especially when the range of the X variable is small. An over-determined model. This happens when the model has more explanatory variables than the number of observations. This could happen in medical research where there may be a small number of patients about whom information is collected on a large number of variables.
3. 4.
An additional reason for multicollinearity, especially in time series data, may be that the regressors included in the model share a common trend, that is, they all increase or decrease over time.
PRACTICAL CONSEQUENCES OF MULTICOLLINEARITY In cases of near or high multicollinearity, one is likely to encounter the following consequences: a) Although BLUE, the OLS estimators have large variances and co variances, making precise estimation
difficult.
b) Because of consequence 1, the confidence intervals tend to be much wider, leading to the acceptance of the
zero null hypothesis (i.e., the true population coefficient is zero) more readily.
c) Also because of consequence 1, the t-ratio of one or more coefficients tends to be statistically insignificant. d) Although the t-ratio of one or more coefficients is statistically insignificant, R 2, the overall measure of
goodness of fit, can be very high.
e) The OLS estimators and their standard errors can be sensitive to small changes in the data.
16 | E C O N O M E T R I C S 2 T E R M P A P E R
STEP: First regress the above equation. Open the File containing data Quick estimate equation write: Y C X1 X2 X3 X4 X5 And check the R and t-ratios.
2
Dependent Variable: Y Method: Least Squares Date: 08/19/13 Time: 13:57 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1 X2 X3 X4 X5
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic
17 | E C O N O M E T R I C S 2 T E R M P A P E R
Durbin-Watson stat
0.196325
Prob(F-statistic)
0.000000
The above regression results shows that R = 0.997288(high).X1, X2, X3, X4 are significant while X5 is insignificant. There may be multicollinearity here.
1.
STEP: First open the File containing data Quick Group Statistics Correlation Write: X1 X2 X3 X4 X5 X1 X1 X2 X3 X4 X5 1.000000 0.898710 0.900580 0.988885 -0.071599 X2 0.898710 1.000000 0.713399 0.863979 -0.222654 X3 0.900580 0.713399 1.000000 0.902906 -0.159905 X4 0.988885 0.863979 0.902906 1.000000 -0.053507 X5 -0.071599 -0.222654 -0.159905 -0.053507 1.000000
REGRESS Y ON X1:
18 | E C O N O M E T R I C S 2 T E R M P A P E R
Dependent Variable: Y Method: Least Squares Date: 08/19/13 Time: 14:02 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1
46.13325 1.20E-07
0.681312 6.11E-09
67.71239 19.57344
0.0000 0.0000
R-squared
2
0.882521
58.49076
0.997288
REGRESS Y ON X2:
Dependent Variable: Y Method: Least Squares Date: 08/19/13 Time: 14:03 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X2
54.61037 7.42E-11
0.659574 8.68E-12
82.79642 8.554942
0.0000 0.0000
19 | E C O N O M E T R I C S 2 T E R M P A P E R
R-squared
2
0.589329
58.49076
0.997288
REGRESS Y ON X3:
Dependent Variable: Y Method: Least Squares Date: 08/19/13 Time: 14:03 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X3
47.82694 3.221181
0.889198 0.245909
53.78662 13.09909
0.0000 0.0000
R-squared
2
0.770875
58.49076
0.997288
REGRESS Y ON X4:
Dependent Variable: Y Method: Least Squares Date: 08/19/13 Time: 14:03 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
20 | E C O N O M E T R I C S 2 T E R M P A P E R
C X4
22.00695 1.241685
1.082998 0.036487
20.32039 34.03118
0.0000 0.0000
R-squared
2
0.957821
58.49076
0.997288
REGRESS Y ON X5:
Dependent Variable: Y Method: Least Squares Date: 08/19/13 Time: 14:04 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X5
58.94157 -0.146728
5.968745 1.927387
9.875036 -0.076128
0.0000 0.9396
R-squared R
2
0.000114 0.997288
58.49076
21 | E C O N O M E T R I C S 2 T E R M P A P E R
REGRESS X1 ON X2 X3 X4 X5:
Dependent Variable: X1 Method: Least Squares Date: 08/19/13 Time: 14:06 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X2 X3 X4 X5
R-squared R
2
0.989170 0.997288
1.03E+08
REGRESS X2 ON X1 X3 X4 X5:
Dependent Variable: X2 Method: Least Squares Date: 08/19/13 Time: 14:07 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
22 | E C O N O M E T R I C S 2 T E R M P A P E R
C X1 X3 X4 X5
R-squared
2
0.910163
5.23E+10
0.997288
REGRESS X3 ON X1 X2 X4 X5:
Dependent Variable: X3 Method: Least Squares Date: 08/19/13 Time: 14:07 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1 X2 X4 X5
R-squared
0.894134
3.310530
23 | E C O N O M E T R I C S 2 T E R M P A P E R
0.997288
REGRESS X4 ON X1 X2 X3 X5:
Dependent Variable: X4 Method: Least Squares Date: 08/19/13 Time: 14:08 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1 X2 X3 X5
R-squared
2
0.981090
29.38251
0.997288
REGRESS X5 ON X1 X2 X3 X4:
Dependent Variable: X5
24 | E C O N O M E T R I C S 2 T E R M P A P E R
Method: Least Squares Date: 08/19/13 Time: 14:08 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1 X2 X3 X4
R-squared
2
0.351577
3.072447
0.997288
The above regression shows data had no multicollinearity. REMEDIAL MEASURES: A. DROPPING A VARIABLE(S) AND SPECIFICATION BIAS :
STEP: First Regress the (10.1) equation without X1 (by assuming it causes Multicollinearity). Open the File containing data Quick estimate equation write: Y C X2 X4 And check the R2 and t-ratios
Dependent Variable: Y
25 | E C O N O M E T R I C S 2 T E R M P A P E R
Method: Least Squares Date: 08/19/13 Time: 14:12 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X2 X3 X4 X5
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
After dropping X1, regression shows that multicollinearity has been removed.
Dependent Variable: D(Y) Method: Least Squares Date: 08/19/13 Time: 14:18
26 | E C O N O M E T R I C S 2 T E R M P A P E R
Variable
Coefficient
Std. Error
t-Statistic
Prob.
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
27 | E C O N O M E T R I C S 2 T E R M P A P E R
THE NATURE OF AUTOCORRELATION The term autocorrelation may be defined as correlation between members of series of observations ordered in time [as in time series data] or space [as in cross-sectional data].In the regression context, the classical linear regression model assumes that such autocorrelation does not exist in the disturbances Ui. Tintner defines autocorrelation as lag correlation of a given series with itself, lagged by a number of time units, whereas he reserves the term serial correlation to lag correlation between two different series.
SOURCES OF AUTOCORRELATION
1. Inertia. A salient feature of most economic time series is inertia, or sluggishness. As is well known, time series such as GNP, price indexes, production, employment, and unemployment exhibit (business) cycles. 2. Specification Bias: Excluded Variables Case 3. Specification Bias: Incorrect Functional Form 4. Cobweb Phenomenon 5. Lags. 6. Manipulation of Data: Another source of manipulation is interpolation or extrapolation of data. 7. Data Transformation
28 | E C O N O M E T R I C S 2 T E R M P A P E R
PRACTICAL CONSEQUENCES OF AUTOCORRELATION In the presence of autocorrelation the usual OLS estimators, although linear, unbiased, and asymptotically (i.e., in large samples) normally distributed, are no longer minimum variance among all linear unbiased estimators. In short, they are not efficient relative to other linear and unbiased estimators. Put differently, they may not be BLUE. As a result, the usual, t, F, and may not be valid.
29 | E C O N O M E T R I C S 2 T E R M P A P E R
It is clear from the Graph that the residuals Rt and Rt-1 are serially correlated and positively correlated
Step 1: Regress the Model Open the File containing data Quick estimate equation Write: Y C X1 X2 X3 X4 X5
Step 2: obtain the Residuals From the estimated equation result window Proc make residuals series (name) ok.
Step 3: check the Runs by looking at changing sign of residuals, then find the interval by calculating mean variance the check whether Runs obtained lies in interval or not for clarity regarding Autocorrelation
30 | E C O N O M E T R I C S 2 T E R M P A P E R
= 1710188/137904 = 12.4012
If the null hypothesis of randomness is sustainable, following the properties of the normal distribution, we should expect that PROB [E(R)-1.96 R E(R) +1.96] PROB [25.653-1.96(12.4012) 8 25.653+1.96(12.4012)] PROB [1.3466 8 49.959]
Obviously, this interval includes 8. So we can accept the null hypothesis that residuals do not contain autocorrelation.
Step 1: Regress the Model Open the File containing data Quick estimate equation Write:
Y C X 2 X 3 X4 X5
In the regression results there is Durbin-Watson d statistics Dependent Variable: Y Method: Least Squares Date: 08/19/13 Time: 13:57 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
31 | E C O N O M E T R I C S 2 T E R M P A P E R
C X1 X2 X3 X4 X5
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
4. A GENERAL TEST OF AUTOCORRELATION: THE BREUSCHGODFREY (BG) TEST MODEL: Yi = 1 + 2 X1i + 3 X2i + 4 X3i +5X4i + 6X5i +i
Step 1: Regress the Model Open the File containing data Quick estimate equation equation Specification window Write:
Y C X 2 X 3 X4 X5
32 | E C O N O M E T R I C S 2 T E R M P A P E R
Step 2: from step 1, in estimated equation window View Residual Tests Serial Correlation LM Test Breusch-Godfrey Serial Correlation LM Test:
F-statistic Obs*R-squared
28.96111 42.88204
Probability Probability
0.000000 0.000000
Test Equation: Dependent Variable: RESID Method: Least Squares Date: 08/19/13 Time: 14:44 Pre sample missing value lagged residuals set to zero.
Variable
Coefficient
Std. Error
t-Statistic
Prob.
-0.511900 -5.69E-09 1.71E-12 0.035050 0.027397 0.030173 0.772369 0.109866 0.067084 -0.015117 0.029055 -0.316267
0.658397 4.60E-09 1.19E-12 0.040369 0.034699 0.062743 0.154615 0.201820 0.201378 0.211764 0.218522 0.171752
-0.777494 -1.238405 1.433777 0.868241 0.789556 0.480893 4.995436 0.544377 0.333126 -0.071384 0.132961 -1.841416
0.4413 0.2226 0.1592 0.3903 0.4343 0.6331 0.0000 0.5891 0.7407 0.9434 0.8949 0.0728
R-squared
0.809095
-2.50E-15
33 | E C O N O M E T R I C S 2 T E R M P A P E R
Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
Step: Regress the Model Open the File containing data Quick estimate equation equation Specification window Write:
YCX2X3
In equation specification window options estimation option window tick Heteroscedasticity-consistent covariance NeweyWest click ok
Dependent Variable: Y Method: Least Squares Date: 08/21/13 Time: 11:22 Sample: 1960 2012 Included observations: 53 Newey-West HAC Standard Errors & Covariance (lag truncation=3)
Variable
Coefficient
Std. Error
t-Statistic
Prob.
34 | E C O N O M E T R I C S 2 T E R M P A P E R
C X1 X2 X3 X4 X5
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
35 | E C O N O M E T R I C S 2 T E R M P A P E R
Write:
Y C X1 X 2 X 3 X4 X5
And take Durbin Watson statistics and calculate rho (= p). Dependent Variable: Y Method: Least Squares Date: 08/19/13 Time: 13:57 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1 X2 X3 X4 X5
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
1 0.196325/2 0.9018
36 | E C O N O M E T R I C S 2 T E R M P A P E R
Step 2: Regress the equation Open the File containing data Quick estimate equation Write: Y- 0.9018*d(Y) 1-0.9018 X1-0.9018*d(X1) X5-0.9018*(X5)
Dependent Variable: Y-0.9018*D(Y) Method: Least Squares Date: 08/21/13 Time: 11:30 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion Durbin-Watson stat
37 | E C O N O M E T R I C S 2 T E R M P A P E R
t = t-1 + t
Step 1: Regress the Model Open the File containing data Quick estimate equation Write:
Y C X1 X 2 X 3 X4 X5
Obtain the Residuals, from the estimated equation result window Proc make Residuals series (name) ok
Step 2: Regress the residual on it lag to get rho to transform model in GDE Write: R1 R1 (-1) Dependent Variable: R1 Method: Least Squares Date: 08/21/13 Time: 11:32 Sample (adjusted): 1962 2012 Included observations: 51 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
D(R1)
0.356889
0.311125
1.147091
0.2568
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion Durbin-Watson stat
38 | E C O N O M E T R I C S 2 T E R M P A P E R
Open the File containing data Quick estimate equation Write: Y- 0.3568*d(Y) 1-0.3568 X1-0.3568*d(X1) X5-0.3568*(X5)
Dependent Variable: Y-.3568*D(Y) Method: Least Squares Date: 08/21/13 Time: 11:35 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion Durbin-Watson stat
from Durbin two steps method: Step 1: Regress the equation given suggested by Durbin for his two step methods Open the File containing data Quick estimate equation
39 | E C O N O M E T R I C S 2 T E R M P A P E R
write: Y C X1 d(X1) X2 d(X2) X3 d(X3) X4 d(X4) X5 d(X5) d(Y) The coefficient of Yt-1 is estimated rho (=P) Dependent Variable: Y Method: Least Squares Date: 08/21/13 Time: 11:39 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
-17.98627 -1.47E-07 -3.78E-07 -2.07E-11 4.08E-12 -0.213495 0.176647 3.177549 -0.389698 -0.037026 0.074601 3.333366
3.065724 1.22E-08 2.41E-07 2.75E-12 8.47E-12 0.091318 0.106567 0.142231 1.727353 0.130887 0.113507 1.153424
-5.866892 -12.00310 -1.570401 -7.517914 0.481797 -2.337915 1.657618 22.34072 -0.225604 -0.282886 0.657232 2.889976
0.0000 0.0000 0.1242 0.0000 0.6326 0.0245 0.1052 0.0000 0.8227 0.7787 0.5148 0.0062
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
40 | E C O N O M E T R I C S 2 T E R M P A P E R
= 3.3333
Step 2: Regress the equation (12.4) Open the File containing data Quick estimate equation Write: Y- 3.3333*d(Y) 1-3.333 X1-3.3333*d(X1) X5-3.3333*(X5)
Dependent Variable: Y-3.333*D(Y) Method: Least Squares Date: 08/21/13 Time: 11:43 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion Durbin-Watson stat
41 | E C O N O M E T R I C S 2 T E R M P A P E R
BerenbluttWebb test g-statistics for estimated =1: To test the hypothesis that 1. The test statistic they use is called the g-statistic, which is defined as follows: g = et2 / t2
Step 1: Regress the equation Open the File containing data Quick estimate equation Write: Y C X1 X2 X3 X4 X5 Obtain residuals, t
2
Step 2: Regress the equation First difference Open the File containing data Quick estimate equation Write: D(Y) C D(X10 D(X2) D (X3) D(X4) D(X5) Obtain residuals, et
2
Dependent Variable: Y Method: Least Squares Date: 08/21/13 Time: 16:27 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1 X2 X3
42 | E C O N O M E T R I C S 2 T E R M P A P E R
X4 X5
2.668870 -0.163192
0.070082 0.129846
38.08228 -1.256813
0.0000 0.2150
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion
Dependent Variable: D(Y) Method: Least Squares Date: 08/21/13 Time: 16:29 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion
g = 0.0311
43 | E C O N O M E T R I C S 2 T E R M P A P E R
It is clear that g statistics lies in the 0 dL range we can take first difference by assuming =1.
THE NATURE OF HETEROSCEDASTICITY A critical assumption of the classical linear regression model is that the disturbances i have all the same variance, 2 if this assumption is not satisfied, there is heteroscedasticity. Hence, there is heteroscedasticity. E i = i2 Notice the subscript of i2, which reminds us that the conditional variances of i (= conditional variances of Yi) are no longer constant
SOURCES OF HETEROSCEDASTICITY
44 | E C O N O M E T R I C S 2 T E R M P A P E R
1. Following the error-learning models, as people learn, their errors of behavior become smaller over time. In this case, i2 is expected to decrease. 2. As data collecting techniques improve, i2 is likely to decrease. 3. Heteroscedasticity can also arise as a result of the presence of outliers. An outlying observation, or outlier, is an observation that is much different (either very small or very large) in relation to the observations in the sample. 4. Another source of heteroscedasticity arises from violating Assumption 9 of CLRM, namely, that the regression model is correctly specified. 5. Another source of heteroscedasticity is skewness in the distribution of one or more regressors included in the model. 6. Other sources of heteroscedasticity: As David Hendry notes, heteroscedasticity can also arise because of (1) incorrect data transformation (e.g., ratio or first difference transformations) and (2) incorrect functional form (e.g., linear versus log linear models).
PRACTICAL CONSEQUENCES OF HETEROSCEDASTICITY In the presence of heteroscedasticity the usual OLS estimators, although linear, unbiased, and asymptotically (i.e., in large samples) normally distributed, are no longer minimum variance among all linear unbiased estimators. In short, they are not efficient relative to other linear and unbiased estimators. Put differently, they may not be BLUE. As a result, the usual, t, F, and may not be valid.
45 | E C O N O M E T R I C S 2 T E R M P A P E R
DETECTION OF HETEROSCEDASTICITY: 1. GRAPHICAL METHOD: MODEL: Yi = 1 + 2 X1i + 3 X2i + 4 X3i +5X4i + 6X5i +i
Step 1: Regress the Model Open the File containing data Quick estimate equation Write: Y C X1 X2 X3 X4 X5
Step 2: obtain the Residuals From the estimated equation result window Proc make residuals series (name) ok.
Step 3: Plot these residuals squares and X1, X2, X3, X4, X5 & individually and see the pattern for Hetroscedasticity. Y From step2 Quick Graph Scatter plot Write: X1 R1^2, X2 R1^2, X3 R1^2, X4 R1^2, X5 R1^2
46 | E C O N O M E T R I C S 2 T E R M P A P E R
It is clear from the graph that there is such pattern seen that causes hetroscedasticity.
2. FORMAL METHODS:
a. PARK TEST:
47 | E C O N O M E T R I C S 2 T E R M P A P E R
Step 1: Regress the Model Open the File containing data Quick estimate equation Write:
Y C X1X 2
And obtain the Residuals from the estimated equation result window Proc make Residuals series (name) ok.
Dependent Variable: Y Method: Least Squares Date: 08/19/13 Time: 13:57 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1 X2 X3 X4 X5
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion
48 | E C O N O M E T R I C S 2 T E R M P A P E R
-7.334900 0.196325
F-statistic Prob(F-statistic)
3456.958 0.000000
Step 2: Regress the Equation (11.1) suggested by Park. Open the File containing data Quick estimate equation Write: log(r1^2) c log(x1) log(r1^2) c log(x2) log(r1^2) c log(x3) log(r1^2) c log(x4) log(r1^2) c log(x5) And check the significance of coefficient of explanatory variable
Dependent Variable: LOG(R1^2) Method: Least Squares Date: 08/21/13 Time: 12:07 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C LOG(X1)
39.65786 -2.361010
10.25252 0.557693
3.868109 -4.233531
0.0003 0.0001
49 | E C O N O M E T R I C S 2 T E R M P A P E R
Dependent Variable: LOG(R1^2) Method: Least Squares Date: 08/21/13 Time: 12:07 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C LOG(X2)
17.17790 -0.866555
5.018112 0.207702
3.423181 -4.172101
0.0012 0.0001
Dependent Variable: LOG(R1^2) Method: Least Squares Date: 08/21/13 Time: 12:07 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C LOG(X3)
-1.586677 -1.935249
0.639137 0.532505
-2.482532 -3.634238
0.0164 0.0007
Dependent Variable: LOG(R1^2) Method: Least Squares Date: 08/21/13 Time: 12:08 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
50 | E C O N O M E T R I C S 2 T E R M P A P E R
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C LOG(X4)
19.37869 -6.848214
5.598606 1.657294
3.461342 -4.132166
0.0011 0.0001
Dependent Variable: LOG(R1^2) Method: Least Squares Date: 08/21/13 Time: 12:08 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C LOG(X5)
-5.142800 1.264550
2.376507 2.120873
-2.164016 0.596240
0.0353 0.5537
If 2s turns out to be statistically significant at 1%, it would suggest that heteroscedasticity is present in the data in the each case.
Step 1: Regress the Model Open the File containing data Quick estimate equation
51 | E C O N O M E T R I C S 2 T E R M P A P E R
Write:
Y C X1X 2
And obtain the Residuals from the estimated equation result window Proc make Residuals series (name) ok
= 1+2+Xi+i
Step 2: Regress the Equation (11.1a) suggested by Glesjer. Open the File containing data Quick estimate equation Write: abs(r1) c x1 abs(r1) c x2 abs(r1) c x3 abs(r1) c x4 abs(r1) c x5 And check the significance of coefficient of explanatory variable
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:16 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1
0.462368 -2.31E-09
0.051704 4.60E-10
8.942531 -5.020498
0.0000 0.0000
52 | E C O N O M E T R I C S 2 T E R M P A P E R
Method: Least Squares Date: 08/21/13 Time: 12:16 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X2
0.290397 -1.30E-12
0.029296 3.82E-13
9.912631 -3.408431
0.0000 0.0013
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:16 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X3
0.423221 -0.060295
0.050627 0.013883
8.359624 -4.343235
0.0000 0.0001
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:17 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
0.914931
0.137492
6.654407
0.0000
53 | E C O N O M E T R I C S 2 T E R M P A P E R
X4
-0.023500
0.004612
-5.094806
0.0000
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:17 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X5
0.123370 0.031866
0.185112 0.059835
0.666459 0.532554
0.5082 0.5967
If 2s turns out to be statistically significant at 1%, it would suggest that heteroscedasticity is present in the data in each case.
= 1+2+Xi+ i
Step 2: Regress the Equation (11.1b) suggested by Glesjer. Open the File containing data Quick estimate equation Write: abs(r1) c x1^0.5 abs(r1) c x2^0.5 abs(r1) c x3^0.5 abs(r1) c x4^0.5
54 | E C O N O M E T R I C S 2 T E R M P A P E R
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:18 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1^.5
0.697759 -4.76E-05
0.093902 9.19E-06
7.430750 -5.181553
0.0000 0.0000
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:18 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X2^.5
0.383647 -8.01E-07
0.041604 1.80E-07
9.221475 -4.443441
0.0000 0.0000
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:18 Sample (adjusted): 1961 2012
55 | E C O N O M E T R I C S 2 T E R M P A P E R
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X3^.5
0.633219 -0.230573
0.090155 0.049249
7.023649 -4.681793
0.0000 0.0000
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:18 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X4^.5
1.599442 -0.254295
0.269902 0.049674
5.926016 -5.119270
0.0000 0.0000
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:19 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X5^.5
0.004676 0.123829
0.366416 0.209160
0.012761 0.592029
0.9899 0.5565
56 | E C O N O M E T R I C S 2 T E R M P A P E R
If 2s turns out to be statistically significant at 1%, it would suggest that heteroscedasticity is present in the data in each case.
= 1+2+1/Xi+ i
Step 2: Regress the Equation (11.1c) suggested by Glesjer. Open the File containing data Quick estimate equation Write: abs(r1) c 1/x1 abs(r1) c 1/x2 abs(r1) c 1/x3 abs(r1) c 1/x4 abs(r1) c 1/x5 And check the significance of coefficient of explanatory variable
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:20 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C 1/X1
-0.010629 20352326
0.048537 3925383.
-0.218995 5.184800
0.8275 0.0000
57 | E C O N O M E T R I C S 2 T E R M P A P E R
Date: 08/21/13 Time: 12:20 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C 1/X2
0.135861 1.42E+09
0.026436 2.98E+08
5.139347 4.761296
0.0000 0.0000
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:20 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C 1/X3
0.009848 0.577436
0.044574 0.110386
0.220930 5.231046
0.8260 0.0000
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:20 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
-0.451516
0.132770
-3.400733
0.0013
58 | E C O N O M E T R I C S 2 T E R M P A P E R
1/X4
19.46190
3.801832
5.119086
0.0000
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:21 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C 1/X5
0.357066 -0.410269
0.180871 0.541472
1.974150 -0.757693
0.0539 0.4522
If 2s turns out to be statistically significant at 1%, it would suggest that heteroscedasticity is present in the data in each case.
= 1+2+1/Xi+ i
Step 2: Regress the Equation (11.1d) suggested by Glesjer. Open the File containing data Quick estimate equation Write: abs(r1) c 1/x1^0.5 abs(r1) c 1/x2^0.5 abs(r1) c 1/x3^0.5 abs(r1) c 1/x4^0.5 abs(r1) c 1/x5^0.5 And check the significance of coefficient of explanatory variable
59 | E C O N O M E T R I C S 2 T E R M P A P E R
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:22 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C 1/X1^0.5
-0.245656 4471.569
0.090575 848.7235
-2.712173 5.268582
0.0091 0.0000
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:22 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C 1/X2^0.5
0.047878 25715.18
0.037971 4894.805
1.260883 5.253565
0.2132 0.0000
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:22
60 | E C O N O M E T R I C S 2 T E R M P A P E R
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C 1/X3^0.5
-0.199286 0.713478
0.083843 0.138597
-2.376886 5.147859
0.0213 0.0000
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/21/13 Time: 12:22 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C 1/X4^0.5
-1.135887 7.318108
0.265139 1.426138
-4.284128 5.131418
0.0001 0.0000
Dependent Variable: ABS(R1) Method: Least Squares Date: 08/22/13 Time: 13:12 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C 1/X5^0.5
0.475984 -0.443667
0.362154 0.629237
1.314313 -0.705088
0.1947 0.4840
61 | E C O N O M E T R I C S 2 T E R M P A P E R
If 2s turns out to be statistically significant at 1%, it would suggest that heteroscedasticity is present in the data in each case
c. WHITES GENERAL HETEROSCEDASTICITY TEST: MODEL: Yi = 1 + 2 X1i + 3 X2i + 4 X3i +5X4i + 6X5i +i
Step 1: Regress the Model Open the File containing data Quick estimate equation Write:
Y C X1X 2 X3 X4 X5
Step 2: applying Whites General Hetroscedasticity Test In estimated equation in step1 view Residuals Tests White Heteroskedasticity Test (cross terms) or White Heteroskedasticity Test (no cross term)
F-statistic Obs*R-squared
5.982381 31.13871
Probability Probability
0.000014 0.000557
As n*R2 (=31.13871) > the critical chi-square (=3.94) value at the 5% level of significance, the conclusion is that there is heteroscedasticity
62 | E C O N O M E T R I C S 2 T E R M P A P E R
F-statistic Obs*R-squared
3.782162 37.24425
Probability Probability
0.000405 0.010937
As n*R2 (=37.24425) > the critical chi-square (=10.85) value at the 5% level of significance, the conclusion is that there is heteroscedasticity
Step 1: Regress the Model Open the File containing data Quick estimate equation Write:
Y C X1X 2 X3 X4 X5
I2= 1+2I2 +i
Step 2: Regress the (11.6a) equation Open the File containing data Quick estimate equation Write: in R1^2 C (Y_CAP)^2 Where I are the estimated values from the model (11.6). The null hypothesis is that 2= 0. If this is not rejected, then one could conclude that there is no heteroscedasticity. The null hypothesis can be tested by the usual t test or the F test,
63 | E C O N O M E T R I C S 2 T E R M P A P E R
Dependent Variable: R1^2 Method: Least Squares Date: 08/21/13 Time: 12:44 Sample (adjusted): 1961 2012 Included observations: 52 after adjustments
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C (Y_CAP)^2
0.414590 -9.74E-05
0.072768 2.07E-05
5.697453 -4.714133
0.0000 0.0000
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
REMEDIAL MEASURES:
WHEN i2 IS KNOWN: THE METHOD OF WEIGHTED LEAST SQUARES: If i2 is known, the most straightforward method of correcting heteroscedasticity is by
Means of weighted least squares, for the estimators thus obtained are BLUE Yi = 1 + 2 X1i + 3 X2i + 4 X3i + 5X4i + 6X5i +i Yi/ = 1 (1/)+ 2 X1i /+ 3 X2i/+ 4 X3i/ +
Where i are the standard deviations
64 | E C O N O M E T R I C S 2 T E R M P A P E R
WHEN i2 IS NOT KNOWN: Whites Heteroscedasticity-Consistent Variances and Standard Errors. MODEL: Yi = 1 + 2 X1i + 3 X2i + 4 X3i +5X4i + 6X5i +i
Step 1: Regress the Model Open the File containing data Quick estimate equation Write:
Y C X1X 2 X3 X4 X5
Dependent Variable: Y Method: Least Squares Date: 08/19/13 Time: 13:57 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1 X2 X3 X4 X5
65 | E C O N O M E T R I C S 2 T E R M P A P E R
Step 2: Regress the Model Open the File containing data Quick estimate equation Write:
Y C X1X 2 X3 X4 X5
In equation specification window options estimation option window tick Heteroscedasticity-consistent covariance Whites click ok
Dependent Variable: Y Method: Least Squares Date: 08/22/13 Time: 14:14 Sample: 1960 2012 Included observations: 53 White Heteroskedasticity-Consistent Standard Errors & Covariance
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1 X2 X3 X4 X5
As we can see that in case of heteroscedasticity the OLS standard errors of slope coefficient of X1 X2 X3 are over estimated and X4 X5 are under estimated. And intercept was under estimated. After applying Whites Hetroscedasticity-Consistent Variances and Standard Errors remedial procedure they were now corrected for hetroscedasticity.
66 | E C O N O M E T R I C S 2 T E R M P A P E R
Step 1: Regress the equation (11.7a) Open the File containing data Quick estimate equation equation specification window Write: Y/X1 C 1/X1 X2/X1 X3/X1 X4/X1 X5/X1
Dependent Variable: Y/X1 Method: Least Squares Date: 08/22/13 Time: 14:31 Sample: 1960 2012 Included observations: 53 White Heteroskedasticity-Consistent Standard Errors & Covariance
Variable
Coefficient
Std. Error
t-Statistic
Prob.
R-squared
0.999641
6.51E-07
67 | E C O N O M E T R I C S 2 T E R M P A P E R
Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
Dependent Variable: Y/X2 Method: Least Squares Date: 08/22/13 Time: 14:33 Sample: 1960 2012 Included observations: 53 White Heteroskedasticity-Consistent Standard Errors & Covariance
Variable
Coefficient
Std. Error
t-Statistic
Prob.
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic
68 | E C O N O M E T R I C S 2 T E R M P A P E R
Durbin-Watson stat
0.158014
Prob(F-statistic)
0.000000
Dependent Variable: Y/X3 Method: Least Squares Date: 08/22/13 Time: 14:34 Sample: 1960 2012 Included observations: 53 White Heteroskedasticity-Consistent Standard Errors & Covariance
Variable
Coefficient
Std. Error
t-Statistic
Prob.
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
69 | E C O N O M E T R I C S 2 T E R M P A P E R
Dependent Variable: Y/X4 Method: Least Squares Date: 08/22/13 Time: 14:35 Sample: 1960 2012 Included observations: 53 White Heteroskedasticity-Consistent Standard Errors & Covariance
Variable
Coefficient
Std. Error
t-Statistic
Prob.
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
Dependent Variable: Y/X5 Method: Least Squares Date: 08/22/13 Time: 14:35 Sample: 1960 2012
70 | E C O N O M E T R I C S 2 T E R M P A P E R
Variable
Coefficient
Std. Error
t-Statistic
Prob.
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
Notice that in the transformed regression the intercept term 2 is the slope coefficient in the original equation and the slope coefficient 1 is the intercept term in the original model. Therefore, to get back to the original model we shall have to multiply the estimated transformed model by Xi.
Assumption 2: The error variance is proportional to Xi. The square root transformation: E (i2) =2Xi
71 | E C O N O M E T R I C S 2 T E R M P A P E R
+6X5i/X1 +i
Step 1: Regress the equation (11.8a) Open the File containing data Quick estimate equation equation specification window Write:
Dependent Variable: Y/X1^0.5 Method: Least Squares Date: 08/22/13 Time: 14:42 Sample: 1960 2012 Included observations: 53 White Heteroskedasticity-Consistent Standard Errors & Covariance
Variable
Coefficient
Std. Error
t-Statistic
Prob.
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
72 | E C O N O M E T R I C S 2 T E R M P A P E R
Similarly do for X2 X3 X4 and X5. Note an important feature of the transformed model: It has no intercept term. Therefore, one will have to use the regression-through-the-origin model to Estimate 1 and 2. Having run (11.8), one can get back to the original model Simply by multiplying transformed model by Xi. Assumption 3: The error variance is proportional to the square of the mean value of Y. E (i2)=2E(Yi)2 Equation postulates that the variance of i2 is proportional to the square of the expected Value of Y. Therefore, if we transform the original equation as follows: Yi /E (Yi) = 1 (1/ E (Yi)) + 2X1/E (Yi) +3 X2i/ E (Yi) + 4 X3i/ E (Yi) + +i +6X5i/ E (Yi)
The transformation (11.9) is, however, in operational because E (Yi) depends on 1 and 2, which are unknown. Of course, we know, which is an estimator of E (Yi) Transforming using: i Yi /i = 1 (1/ i) + 2X1i/ i + 3 X2i/ i + 4 X3i/ i + +6X5i/ i + i
Step: Regress the equation (11.8a) Open the File containing data Quick estimate equation equation specifications window Write: Y/Y_CAP 1/Y_CAP X1/Y_CAP X2/Y_CAP X3/Y_CAP X4/Y_CAP X5/Y_CAP
Dependent Variable: Y/Y_CAP Method: Least Squares Date: 08/22/13 Time: 14:55 Sample: 1960 2012 Included observations: 53
73 | E C O N O M E T R I C S 2 T E R M P A P E R
Variable
Coefficient
Std. Error
t-Statistic
Prob.
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion Durbin-Watson stat
Assumption 4: A log transformation such as: lnYi = 1 + 2 lnXi+ i very often reduces heteroscedasticity. Log Transforming: lnYi = 1 + 2 lnX1i + 3 lnX2i + 4 lnX3i + 5lnX4i+6lnX5i+i
Step: Regress the equation (11.8c) Open the File containing data Quick estimate equation equation specifications window Write: log(y) c log(x1) log(x2) log(x3) log(x4) log(x5)
74 | E C O N O M E T R I C S 2 T E R M P A P E R
Dependent Variable: LOG(Y) Method: Least Squares Date: 08/22/13 Time: 15:00 Sample: 1960 2012 Included observations: 53 White Heteroskedasticity-Consistent Standard Errors & Covariance
Variable
Coefficient
Std. Error
t-Statistic
Prob.
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
This result arises because log transformation compresses the scales in which the variables are measured, thereby reducing a tenfold difference between two values to a twofold difference. To conclude our discussion of the remedial measures, we reemphasize that all the transformations discussed previously are ad hoc; we are essentially speculating
75 | E C O N O M E T R I C S 2 T E R M P A P E R
about the nature of i2. Which of the transformations discussed previously will work will depend on the nature of the problem and the severity of heteroscedasticity.
THE NATURE OF SIMULTANEOUS-EQUATION MODELS In contrast to single-equation models, in simultaneous-equation models more than one dependent, or endogenous, variable is involved, necessitating as many equations as the number of endogenous variables. A unique feature of simultaneous-equation models is that the endogenous variable (i.e., regressand) in one equation may appear as an explanatory variable (i.e., regressor) in another equation of the system. Y1i = 10 + 12Y2i + 11 X1i + u1i Y2i = 20 + 21Y1i + 21 X1i + u2i Where Y1 and Y2 are mutually dependent, or endogenous, variables and X1 is an exogenous variable and where u1 andu2 are the stochastic disturbance terms, the variables Y1 and Y2 are both stochastic. Therefore, unless it can be shown that the stochastic explanatory variable Y2 in (18.1) is distributed independently of u1 and the stochastic explanatory variable Y1in (18.2) is distributed independently of u2, application of the classical OLS to these equations individually will lead to inconsistent estimates.
76 | E C O N O M E T R I C S 2 T E R M P A P E R
As a consequence, such an endogenous explanatory variable becomes stochastic and is usually correlated with the disturbance term of the equation in which it appears as an explanatory variable. In this situation the classical OLS method may not be applied because the estimators thus obtained are not consistent, that is, they do not converge to their true population values no matter how large the sample size. Since simultaneous-equation models are used frequently, especially in econometric models, alternative estimating techniques have been developed by various authors.
Where Yi and X1i are mutually dependent, or endogenous, variables X2i, X3i, X4i and X5i are an exogenous variable and where 1i and 2i are the stochastic disturbance terms, the variables Yi and X1i are both stochastic. Therefore, unless it can be shown that the stochastic explanatory variable Yi is distributed independently of 1i and the stochastic explanatory variable X1i is distributed independently of 2i, application of the classical OLS to these equations individually will lead to inconsistent estimates.
THE IDENTIFICATION PROBLEM RULES FOR IDENTIFICATION THE ORDER CONDITION OF IDENTIFIABILITY A necessary (but not sufficient) condition of identification, known as the order condition, May be stated in two different but equivalent ways as follows (the necessary as well as
77 | E C O N O M E T R I C S 2 T E R M P A P E R
MKG-1
M = numbers of variables in the simultaneous model
i
Hence equation 1 is just identified and equation 2 is over identified. As we know equation 1 is just identified, is estimated using ILS and 2SLS and equation 2 is over identified, is estimated using 2SLS method.
Here we can not apply the rank condition there are only two simultaneous equations only
on X1 X2 X3 X4 X5
78 | E C O N O M E T R I C S 2 T E R M P A P E R
Write: Y C X1 X2 X4 X5 And obtain the Residuals from the estimated equation result window Proc make Residuals series (name) ok
Dependent Variable: Y Method: Least Squares Date: 09/04/13 Time: 12:58 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X2 X3 X4 X5
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
79 | E C O N O M E T R I C S 2 T E R M P A P E R
Open the File containing data Quick estimate equation Write: X1 C Y X2 X3 R1 Now check the significance of this residual coefficient if it exceeds critical then we can say both the equation are simultaneous
Dependent Variable: X1 Method: Least Squares Date: 09/04/13 Time: 13:01 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C Y X2 X3 RESID01
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
As residual is significant 1% level of significance; w can conclude that both equation are simultaneous
80 | E C O N O M E T R I C S 2 T E R M P A P E R
Dependent Variable: Y Method: Least Squares Date: 09/04/13 Time: 13:05 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X1 X2 X4 X5 RESID01
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic
81 | E C O N O M E T R I C S 2 T E R M P A P E R
Durbin-Watson stat
0.248237
Prob(F-statistic)
0.000000
As residual is significant 1% level of significance; w can conclude that both equation are simultaneous
THE NATURE OF DUMMY VARIABLE Now, we consider models that may involve not only ratio scale variables but also nominal scale variables. Such variables are also known as indicator variables, categorical variables, qualitative variables, or dummy variables. Since dummy usually indicate the presence or absence of a quality or an attribute, such as male or female, black or white, Catholic or non-Catholic, Democrat or Republican, they are essentially nominal scale variables. Variables that assume
82 | E C O N O M E T R I C S 2 T E R M P A P E R
such 0 and 1 values are called dummy variables. Dummy variables can be incorporated in regression models just as easily as quantitative variables. As a regression model may contain regressors that are all exclusively dummy, or qualitative, in nature. Such models are called Analysis of Variance (ANOVA) models.
1. For each qualitative regressor the number of dummy variables introduced must be one less than the categories of that variable. If you do not follow this rule, you will fall into what is called the dummy variable trap, that is, the situation of perfect co linearity or perfect multi-co linearity, if there is more than one exact relationship among the variables. 2. The category for which no dummy variable is assigned is known as the base, benchmark, control, comparison, reference, or omitted category. And all comparisons are made in relation to the benchmark category. 3. The intercept value (1) represents the mean value of the benchmark category. In Equation (9.1), the benchmark category is the Rural-North region. 4. The coefficients attached to the dummy variables in (9.1) are known as the differential intercept coefficients because they tell by how much the value of the intercept that receives the value of 1 differs from the intercept coefficient of the benchmark category. Which is a better method of introducing a dummy variable: (1) intro-duce a dummy for each category and omit the intercept term or (2) include the intercept term and introduce only (m1) dummies, where m is the number of categories of the dummy variable? Regression models containing an admixture of quantitative and qualitative variables are called analysis of covariance (ANCOVA) models. ANCOVA models are an extension of the ANOVA models in that they provide a method of statistically controlling the effects of quantitative regressors, called covariates or control variables, in a model that includes both quantitative and qualitative, or dummy, regressors.
83 | E C O N O M E T R I C S 2 T E R M P A P E R
ANALYSIS OF VARIANCE (ANOVA) MODEL: MODEL: Yi = 1 + 2D2i + 1i Where Yi is for life expectancy D2 = 0 for male D2 = 1 for female
First Regress the (9.1) equation Open the File containing data Quick estimate equation Write: y c d2
Dependent Variable: Y Method: Least Squares Date: 09/04/13 Time: 13:19 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C D2
58.44163 0.100143
1.046867 1.494661
55.82529 0.067001
0.0000 0.9468
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
84 | E C O N O M E T R I C S 2 T E R M P A P E R
When D2=0 Yi = 1 + 2D2i + 1i Yi = 58.44163 + 0.100143(0) + 1i Yi = 58.44163 mean life expectancy for male
When D2 = 1 Yi = 1 + 2D2i + 1i Yi = 58.44163 + 0.100143(1) + 1i Yi = 58.541773 mean life expectancy for female
ANALYSIS OF COVARIANCE (ANCOVA) MODEL MODEL: Yi = 1 + 2X3i + 3D2i + 1i Where Yi is life expectancy X3i is unemployment
D2 = 1 for Urban D2 = 0 for Other (Rural)
First regress the equation Open the File containing data Quick estimate equation Write: Y C X3 D2
85 | E C O N O M E T R I C S 2 T E R M P A P E R
Dependent Variable: Y Method: Least Squares Date: 09/04/13 Time: 13:35 Sample: 1960 2012 Included observations: 53
Variable
Coefficient
Std. Error
t-Statistic
Prob.
C X3 D2
R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood Durbin-Watson stat
Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion F-statistic Prob(F-statistic)
When D2 = 0 Yi = 47.80963+ 3.221096X3i+ 1i mean life expectancy function for rural When D2 = 1 Yi = 47.80963 + 3.221096X3i + 0.035865(1) + 1i Yi = 47.845495 + 3.221096X3i + 1i mean life expectancy function for urban