Академический Документы
Профессиональный Документы
Культура Документы
Eunice Han
ehan@fas.harvard.edu
March 26th, 2013 Harvard University
Information
2
Topic
3
OLS
1.
The Assumptions
Omitted Variable Bias
Conditional Mean Independence
Hypothesis Testing and Confidence Intervals
Homoskedasticity vs. Heteroskedasticity
Nonlinear Regression Models: Polynomials, Log Transformation,
and Interaction Terms
2.
Panel Data
3.
Topic (Cont.)
4
4.
5.
6.
Stationarity
Forecasting Models: AR and ADL Model
Dynamic Causal Effects: Distributed Lag Model
Serial Correlation and Newey-West HAC SE
Perfect Multicolinearity
6
Yi 1X1i 2 X 2i i
X
Yi
2 2i
i
)
E(
2
2
1
Corr(X1i , X 2i )
Var(X 2i )
Measure of Fit
11
Hypothesis Testing
12
1
)
Var(
1
~ N(0,1)
Pr(| z || t |) 2( | t |)
Confidence Interval
14
Yi 0 1 X i 2 X i2 3 X i3 ui
Case
Linear-Log
Log-Linear
Log-Log
Regression Specification
Yi 0 1 ln(X i ) ui
ln(Yi ) 0 1X i ui
1
1% change in X 0.01 1
Interpretation of
change in Y
1 unit changein X100 1 %
change in Y
1% change in X 1 %
ln(Yi ) 0 1 ln(X i ) ui change in Y
The interactions
Consider regression Yi 0 1X1i2 X 2i ui
If we believe that the effect of X2i on Y depends on X1i,
we can include the interaction term X1i*X2i as a
regressor;
Yi 0 1X1i2 X 2i 3 (X1i * X 2i ) ui
Interaction terms
20
Yi 1 X 1i 2 X 2i k X ki ui
E (Yi | X i ) Pr(Yi 1 | X 1i ,..., X ki ) 0 1 X 1i ... k X ki
Advantage:
Simple
Disadvantages:
predicted probabilities can be < 0or >1, which
does not make sense in probability.
This can be solved by introducing nonlinear probability
model: probit and logit regression
The
1
E(Yi | X i ) Pr(Yi 1 | X i ) F(0 1 X1)
1 exp (0 1 X1 )
Pr(Yi 1| X) [cutoff
1 (0 1X i )]
Pr(Y
i 2 | X) [cutoff 2 (0 1X i )] [cutoff 1 (0 1X i )]
Pr(Y
(0 1X i )]
i 3 | X) 1 [cutoff 2
Types of Data
30
Panel Data
31
Yit 0 1X1it ... k Xkit (1D1 ... N 1DN 1) (1P1 ... T 1PT 1) uit
Differences-in-Differences Estimator
34
Consider regression Yi 0 1X i ui
Suppose E(u|X) 0; X is endogenous.
IV regression breaks X into two parts by using the
instrumentalvariable Z
1. a part that is correlated with u
2. a part that is not correlated with u (this part is
correlated with Z).
By focusing on the variation in X that is not correlated with
u using the help from the instrument Z, it is possible to
obtain an unbiased and consistent estimate of 1 .
1.
2.
TSLS (cont.)
41
Var( X i ) (Z, X)
where is the sample covariance, is the coefficient
of Z in the regression of Y on Z, and is the coefficient
of Z in the regression of X on Z.
2.
Over-identified
if m > k
The number of instruments > the number of endogenous
regressors.
3.
Under-identified if m < k
The number of instruments < the number of endogenous
regressors. In this case, you need to find more instruments.
Step 1: Run the first stage regression with all Zs and Ws;
X i 0 1Z1i 2Z2i ... m Zmi m1W1i ... mrWri vi
Step 2: Run the joint hypothesis to test H0: 1 2 ... m 0
Step 3: The instruments are relevant if at least one of the m
coefficients is nonzero. If the F-statistic >10, reject the null
and conclude that our instrument(s)
are highly relevant. If
u TSLS
i
u TSLS
0 1Z1i ... mZmi m1W1i ... mrWri i
i
1i 2i ... mi 0
J ~ m2 k
Data for the same entity observed over time. There is one
observation at one point in time.
When we analyze time series data, we are mainly
interested in impulse(shock)-response relationship. For
example, we can examine the impact of rise in oil price on
stock market index or the impact of monetary policy on
aggregate demand.
The purpose in time series study is often to make a good
forecast.
The first lag of Yt is Yt-1, the second lag is Yt-2, etc.
The correlation of a series with its own lagged values is
called autocorrelation (or serial correlation).
Stationarity
46
Model Selection
49
ln
(
p
1
)
Bayesian Information Criterion:
T
T
First term: Always decreasing in p (larger p, better fit)
Second term: Always increasing in p. (Penalty term for
increased forecast error due to increased p)
Choose p that minimizes BIC.
Yt 0 1X t 2 X t 1 3 X t 2 ... r1X t r ut
Newey-West HAC SE
52