Академический Документы
Профессиональный Документы
Культура Документы
2019
1 / 38
Recap
2 / 38
I Under these assumptions, the variance-covariance matrix of βb
conditional on X is σ 2 (X0 X)−1 , which we estimate using
Var\
(β̂ | X) = σ̂ 2 (X0 X)−1
where Pn
ûi2 û0 û
σ̂ 2 = i=1
=
n−k −1 n−k −1
I The standard errors of coefficients reported by all statistical
packages are the square root of the diagonal elements of this
estimated variance-covariance matrix
I Statistical packages also report the standard error of the regression,
which is σ̂, the square root of σ̂ 2
3 / 38
Lecture Outline
I Sampling Distributions of the OLS Estimators (Textbook reference
4-1)
I Testing Hypotheses About a Single Population Parameter
(Textbook reference 4-2)
I Confidence Intervals (Textbook reference 4-3)
I Testing Single Linear Restrictions (Textbook reference 4-4)
I Testing a joint hypothesis of several restrictions - The F test
I An important special case: Testing the overall significance of a
regression model
I Application of t-test to testing a single linear restriction involving
several parameters (textbook reference 4.4)
4 / 38
Sampling Distributions of the OLS Estimators
5 / 38
I Assumptions we have made so far gave us:
E (βb | X) = β
and
Var (βb | X) = σ 2 (X0 X)−1
6 / 38
I Assumption MLR.6 or E.5 (Normality): Conditional on X the
population errors are normally distributed
I The random sampling assumption implies that population errors are
independent of each other
I The zero conditional expectation assumption implies that
population errors have mean zero
I The no heteroskedasticity assumption implies that population errors
have variance σ 2
I Assumption MLR.6 can therefore be expressed as: conditional on
explanatory variables, population errors ui are i.i.d.N(0, σ 2 )
I Similarly, E.5 can be expressed as u | X ∼ N(0, σ 2 In )
I Assumptions MLR.1 to MLR.6 are called the Classical Linear Model
(CLM) assumptions
7 / 38
The Classical Linear Model
f (y | x)
β0 + β1 x
x1
x2 x
x3
8 / 38
I The beauty of Normal distribution is that any linear combination of
Normal random variable is also Normally distributed
I Recall that:
βb = (X0 X)−1 X0 y
= β + (X0 X)−1 X0 u
9 / 38
I That is for j = 0, 1, . . . , k
β̂j | X ∼ N βj , Var (β̂j )
where
Var (β̂j ) = σ 2 (X0 X)−1 jj
β̂j − βj
∼ N(0, 1)
sd(β̂j )
because ...
10 / 38
Testing Hypotheses About a Single Population Parameter
β̂j − βj
∼ N(0, 1)
sd(β̂j )
11 / 38
I Under the CLM Assumptions,
β̂j − βj
∼ tn−k−1 = tdf
se(β̂j )
(no proof required)
I The t distribution also has a bell shape, but is more spread out (has
fatter tails) than the N(0, 1). It gets more similar to N(0, 1) as its
df increases.
12 / 38
I We use the result on the t distribution to test the null hypothesis
about a single βj
I Most routinely we use it to test if controlling for all other x, xj has
no partial effect on y :
H0 : βj = 0
for which we use the t statistic (or t ratio),
β̂j
tβ̂j =
se(β̂j )
13 / 38
I The alternative hypothesis can be one-sided, such as H1 : βj < 0 or
or H1 : βj > 0, or it can be two-sided H1 : βj 6= 0.
I The alternative determines what kind of evidence is considered
legitimate against the null. For example, for H1 : βj < 0 we only are
excited to reject the null if we find evidence that it is negative. We
won’t be interested in any evidence that βj is positive, and we won’t
reject the null if we found such evidence.
I Example: The effect of missing lectures on final performance, the
alternative hypothesis is that missing lectures has negative effect on
your final, even after controlling for how smart you are. We are not
interested in any evidence that missing lectures improves final
performance
I With two sided alternatives, we take any evidence that βj may not
be zero, whether positive or negative, as legitimate
14 / 38
I We also need to specify α the size or the significance level of the
test, which is the probability that we wrongly reject the null when it
is true (Type I error)
I Using the tn−k−1 distribution, the significance level and the type of
alternative, we determine the critical value that defines the rejection
region.
I For H1 : βj > 0
15 / 38
I For H1 : βj 6= 0
I If tcalc (the value of the t-statistic in our sample) falls in the critical
region, we reject the null
I When we reject the null, we say that xj is statistically significant at
the α% level
I When we fail to reject the null, we say that xj is not statistically
significant at the α% level
16 / 38
I We can use t test to test the null hypothesis:
H0 : β j = r
β̂j − r
∼ tn−k−1 if H0 is true
se(β̂j )
17 / 38
Steps involved in statistical verification of a question
18 / 38
Confidence Intervals
β̂j ± c × se(β̂j )
19 / 38
p-value of a test
20 / 38
Example: Effect of missing classes on the final score
21 / 38
Testing multiple linear restrictions: The F test
y = β0 + β1 x 1 + β2 x 2 + β3 x 3 + β 4 x 4 + u
H 0 : β3 = β4 = 0
H0 : β1 = −β2 , β3 = β4 = 0
22 / 38
I The first null involves four restrictions, the second involves ...
restriction and the third involves ... restrictions
I The alternative can only be that at least one of these restrictions is
not true
I The test statistic involves estimating two equations, one without
restrictions (the unrestricted model) and one with the restrictions
imposed (the restricted model), and seeing how much their sum of
squared residuals differ
I This is particularly easy for testing exclusion restrictions like the first
two nulls on the previous slide
23 / 38
I For example for
H0 : β3 = β4 = 0 (note: 2 restrictions)
the alternative is
y = β0 + β1 x1 + β2 x2 + β3 x3 + β4 x4 + u (UR)
y = β0 + β 1 x 1 + β2 x 2 + u (R)
24 / 38
I The test statistic is
(SSRr − SSRur )/q
F = ∼ Fq,n−k−1 under H0
SSRur /(n − k − 1)
25 / 38
I Suppose q = 3 and n − k − 1 = 60. Then from Table G.3b the 5%
Fcrit = 2.76
I We reject the null if Fcalc > 2.76, and not reject otherwise
26 / 38
A useful formulation of the F test
I The SST of the restricted and the unrestricted models are the same
(why?), therefore we have:
27 / 38
Test for overall significance of a model
I This is the F -statistic that is reported in the eviews output any time
we estimate a regression
I It is for the special null hypothesis that all slope parameters are zero
H 0 = β1 = β2 = · · · = βk = 0
28 / 38
Example
I Determinants of long-term economic growth: Controlling for initial
conditions (the size of GDP in 1960 and human capital in 1960)
and openness is the rule of law a significant determinant of long
term growth rate?
I Data from “Finance and the Sources of Growth” Journal of
Financial Economics, Vol. 58, pp. 261-300, published in 2000.
I The unrestricted model:
29 / 38
I H0 : βrev coups = βassasinations = 0
I H1 : at least one of them is not zero
(SSRr −SSRur )/2
I F = SSRur /(65−5−1) ∼ F2,59 under H0
I The restricted model:
(157.1668−147.6766)/2
I Fcalc = 147.6766/(65−5−1) = 1.896 in this sample
30 / 38
I Level of significance of the test: 5%
I Critical value from Table G.3b between 3.15 for F2,60 and 3.23 for
F2,40 . In such cases we prefer to be conservative and choose the
closest degree of freedom less than what we are after
I Fcalc < 3.23 ⇒ There is not enough evidence to suggest that
revolution, coup d’etat or political assassinations have any effect on
long-term growth, after controlling for the effects of initial gdp per
capita, educatedness of the workforce and openness to trade.
I The F-test for overall significance of both regressions shows that
both models have at least one significant explanatory variable for
explaining growth.
31 / 38
Testing a single restriction that involves more than one
parameter
Textbook reference 4.4
y = β0 + β1 x 1 + β2 x 2 + β3 x 3 + u
H 0 : β1 = β2 (Null1)
H 0 : β1 + β2 + β3 = 1 (Null2)
32 / 38
I We can use the F test as usual, there would be no good reason to
study this case separately
I However, with a single hypothesis, we can have one-sided
alternatives. For example for Null1, we can have
H1 : β1 > β2 (Alt1)
H 1 : β 1 + β2 + β3 < 1 (Alt2)
33 / 38
I Consider Null1. Define
δ = β1 − β2
I We can write Null1 and Alt1 as
H0 : δ=0
H1 : δ>0
34 / 38
Reparameterisation: A cool useful trick
H0 : β1 − β2 = 0
H1 : β1 − β2 > 0
and set
δ = β1 − β2 ⇒ β1 = δ + β2
I Replace β1 in the model with δ + β2 and rearrange
y = β0 + (δ + β2 )x1 + β2 x2 + β3 x3 + u ⇒
y = β0 + δx1 + β2 (x1 + x2 ) + β3 x3 + u
35 / 38
Reparameterisation: Example
I Test that controlling for IQ, one additional year of education has the
same proportional effect on wage as one additional year of
experience, against the alternative that it has a larger effect:
37 / 38
Summary
I Using the significance level of the test, we can find the Fcrit of the
test from the F table
I We will learn in tutorial how to get this from eviews
I We reject the null of Fcalc > Fcrit , and not reject otherwise
I The F-test for overall significance of a regression model is an
important case that is provided by all statistical packages
I Finally, we learned how to test general linear restrictions
38 / 38