Академический Документы
Профессиональный Документы
Культура Документы
1st Lecture
Roman Horvath
Roman.horvath@gmail.com
http://ies.fsv.cuni.cz/en/staff/horvath
Organization
Summer semester, every Thursday
5PM lecture (room 314), 6.20PM seminar 1, 9.30AM seminar 2, 11.00AM seminar 3, 12.20PM seminar 4
Lecturers:
Jozef Barunk (barunik@utia.cas.cz)
Jaromir Baxa (jaromir.baxa@centrum.cz)
Roman Horvath (roman.horvath@gmail.com)
TAs:
Frantisek Cech (fero.cech@gmail.com)
Jiri Kukacka (bjiri.kukacka@gmail.com)
Marek Rusnak (rusnakmarek@seznam.cz)
MA course, Elective course, 6 credits
Course website
There is course website, where all materials such
as lecture notes, data or futher readings are
stored:
http://ies.fsv.cuni.cz/en/syllab/JEM116
Exam:
60% for the final test
40% for the term paper
Syllabus
1. Introduction
Readings
Mandatory are lecture notes (ppt presentations that you will see during
lectures)
Several applied econometrics textbooks are recommended
Enders, W.: Applied Econometric Time Series, 2nd edition, 2003
Harris, R. and R. Sollis: "Applied Time Series Modelling and
Forecasting", 2003
Stewart, K. G.: "Introduction to Applied Econometrics", 2005
Verbeek, M.: A Guide to Modern Econometrics, 2nd edition, 2004
Kratzig, M. and H. Lutkepohl ,Applied Time Series Econometrics,
2004
Kocenda, E. and A. Cerny, "Elements of Time Series Econometrics",
2007, Karolinum
Other suggested readings include journal articles (see course website for the
full list)
Assessment
- written exam at the end of semester (60%)
- term paper (40%)
Free software
Many partial packages
In my opinion, the most comprehensive is
JMulTi, Gretl, and may be EasyReg
Great advantage: dowloadable for free, small
disadvantage: help and user guide does not have
to have 1000 well-structured pages as for the Eviews (but typically help at the sufficient level,
see JMulTi)
JMulTi + Gretl
JMulTi downloadable from www.jmulti.com
Gretl downloadable from
http://gretl.sourceforge.net/
Online help + help in pdf documents
Gretl and JMulTi will be introduced during the
seminars
What is it?
How to estimate it?
Autocorrelation
Heteroscedasticity
Multicollinearity
Omitted variable bias
Correct functional form
Classical assumptions
(sometimes too restrictive)
OLS continued
The fit of the model measured by the coefficient of
determination: R-squared (value between zero and one)
As the value of R-squared is positively associated with
the number of explanatory variables, you may use
adjusted R-squared
R-squared adjusted for the number of explanatory variables
imposes punishment for increasing the number of expl. variables
OLS continued
The estimated parameters have their standard
errors
You may form the t-statistic, if you want to see that
estimated parameter is significantly different from zero,
just divide the parameter by its standard error, you may
also be interested in p-value (exact significance), if pvalue is smaller than 0.05, then we say the variable is
significant at 5% significance level (t-stat greater than 1.96
in other words)
Multicollinearity
Perfect multicollinearity is where there is exact
linear relationship between explanatory variables
included in the regression
Perfect multicollinearity never happens in
practice, unless you fall in a dummy variable trap
For example, you include both the dummy for large and
small firms and the constant in your regression
Principal components
Form linear combination of correlated variables such that it maximizes its
variability
Drawback: may lack economic reasoning
Omitted variables
If you omit important explanatory variable, the
estimates are biased, and also inconsistent (i.e. bias does
not disappear, if you increase sample size)
Standard errors biased no inference valid
You are then trying to force the regression line where it
does not want to go!
If you include irrelevant variable, estimates unbiased,
but standard errors greater than necessary
Low R-squared
Insignificant variables
Wrong signs
Bad Durbin-Watson values
1.
2.
3.
Heteroscedasticity
Consequences of heteroscedasticity
Still unbiased coefficients, but the standard
errors not correctly estimated so significance
tests invalid
Predictions inefficient
Coefficient of determination not valid
Important to test for any heteroscedasticity and
remedy it if present
Autocorrelation
More likely in time-series
Autocorrelation means that there is some kind
of relationship between the t-th and t-i-th error
term
Say if error term is large in time t, it is very likely
that it will remain high next period, but classical
assumptions want no relationship!
Consequences of autocorrelation
Identical as for heteroscedasticity, standard
errors affected invalid inference
Formal tests:
Durbin Watson test for AR(1)
D-W test statistic roughly equal to 2*(1-P)
P is correlation between the error term and its 1st lag
D-W statistic symmetrically distributed around 2, if it is far from
2, there is autocorrelation, there is a grey zone, where you cant
conclude about autocorrelation based on D-W test, negative
autocorrelation if test statistic enough larger than 2, opposite
for positive autocorrelation
Critical values found in most textbooks, the values depend on
sample size and number of explanatory variables
Remedial measures
First differencing
Quasi differencing
Newey-West wash, analogy of White-washing,
it makes your standard errors and covariance
robust to unknown form of autocorrelation,
available for E-views
Parsimony
Goodness of fit
Theoretical consistency has to make sense
Predictive power