Econometrics Review
Econometrics Review
Econometric Methods
A Quick Review
Francesco Cinnirella
Tractability of variables: A lot of effort has been put into collecting and compiling
new variables based on original or secondary historical sources
Instrumental Variables
Introduction
Introduction
The instrumental variable (IV) estimator is very often used in applied research.
Advantage: with the IV estimator, we can obtain consistent estimates with any
type of correlation between the error term and the explanatory
variables.
Disadvantage: an appropriate instrument needs to be found.
Introduction
Intuition
The OLS estimator exploits the total variation of the dependent and independent
variables.
Introduction
Validity
Y = β0 + β1 X + u (2)
Introduction
Validity
1 The first stage decomposes X into two components: a problematic component that
may be correlated with the regression error, and another problem-free component
that is uncorrelated with the error.
The first stage begins with a population regression linking X and Z (reduced form
equation):
Xi = π0 + π1 Zi + vi (3)
The first stage of 2SLS uses the predicted value from the OLS regression,
X̂i = π̂0 + π̂1 Zi .
The second stage of 2SLS is easy: regress Yi on X̂i using OLS. The resulting
estimators from the second stage regression are the 2SLS estimators.
Yi = β0 + β1 X̂i + ui (4)
where
ln(wage): natural logarithm of wage,
educ: years of education,
exp: years of working experience.
It is plausible to assume that u is uncorrelated both with exp and with exp2 .
We also believe (for the moment, at least) that mother’s and father’s education
(meduc and f educ) are uncorrelated with u.
First, we estimate (5) by OLS (we also want to compare OLS and 2SLS results):
Therefore:
β
c1
OLS = 0.107 , versus β1 2SLS = 0.061
c
(0.014) (0.031)
The correlation between the instrument(s) and the endogenous variable can be tested
statistically.
X = γ0 + γ1 Z + v (9)
If, together with X, the model contains further explanatory variables, they all have to
be included in (9) as well.
If we have two instrumental variables (Z1 and Z2 ) for X, instead of (9), we estimate
the regression
X = γ0 + γ1 Z1 + γ2 Z2 + v (10)
and then test the hypothesis γ1 = γ2 = 0 with an F -test.
In a model with one endogenous explanatory variable and one instrument, assumption
IV2 cannot be statistically tested.
The reason being that the test would involve a correlation between the instrumental
variable and an unobservable error.
If several instruments are available for one single endogenous explanatory variable, we
can test the overidentifying restrictions (but in this case we have a remaining
non-testable core of assumption IV2).
Then we reject H0 if nR2 is larger than the critical value of the χ2q of our choice.
Assumption IV1 states that instruments and endogenous variable must be correlated.
A correlation equal to 0 is improbable in practice.
But a serious problem can arise if the correlation between the instrument and the
endogenous variable is small.
In this case the IV or 2SLS estimators cannot provide reliable results. In the literature,
this is called the problem of weak instruments.
A serious problem can arise if the correlation between the instrument and the
endogenous variable is small.
In this case the IV or 2SLS estimators cannot provide reliable results. This is
called the problem of weak instruments.
If the relationship between the instrument and the explanatory variable is weak,
but assumption IV2 is satisfied, then:
the bias of the IV estimator can be very large, even in large samples;.
the IV estimator is not normally distributed, even in large samples.
In the case of one endogenous explanatory variable and several instruments, the
following holds as “rule of thumb”: in the first-stage regression, if the F -statistic
for overall significance of the instrument(s) is larger than 10, the instruments are
not considered as being weak.
In the case that assumption IV2 is violated, then the IV estimator is always
inconsistent.
Natural experiments
We will first deal with the concept of true (“ideal”) randomised experiment.
The causal effect of a variable (e.g., a policy intervention) on other variables can be
reliably estimated using a true randomised experiment.
Experiments
Terminology: experiments and quasi-experiments
For example, we want to estimate the effect of a training programme for unemployed.
For this purpose, we randomly choose a group of unemployed for the programme and
we compare their outcome after the programme with a group of unemployed
individuals who did not take part to the program.
After the experiment we can evaluate the effect of the programme simply comparing
the treatment and control groups.
the IV estimator
For example, we observe the outcome of a group of people both before and after
an experiment
The sample is broken down into four groups: the control group before and after the
treatment, and the treatment group before and after the treatment.
where:
treat is a dummy variable taking value 1 if the person belongs to the treatment
group.
af ter is a dummy variable taking value 1 if the observation takes place after the
treatment.
βbDiD is also called the average treatment effect (ATE), because it measures the
effect of the “treatment” on the average outcome Y .
Identifying assumption: the treatment units have similar trend to the control units in
the absence of treatment
In an experiment with partial compliance, the assigned treatment level can serve
as an instrumental variable for the actual treatment level.
Relevance: as long as the protocol is partially followed, the actual treatment level
is partially determined by the assigned treatment level (instrument is relevant).
Natural Experiments
Natural experiments occur when some exogenous event (usually a policy change)
changes the environment where subjects operate.
Natural Experiments
The econometric methods used are similar to those outlined for true experiments.
In order to control for systematic differences between the control and treatment
groups, we need at least two years of data.
Often panel data is not available, therefore we need to use repeated cross-sections.
Conclusion
When the “as if” randomization only partly influences the treatment, then
instrumental variables regression can be used instead.
In addition, the fact that the assignment to the treatment and control groups
follows a non-linear pattern (the discontinuity at exactly the cut-off value) allows
to control for any smooth function of the variable determining eligibility.
Assumption for causal effect: there are no other discontinuities around the
cut-off.
The next figure illustrates a hypothetical RD scenario where those with xi ≥ 0.5
are treated.
In both cases, there is a discontinuity in the relation between E[Y0i |xi ] and xi
around the point x0 .
0 .2 .4 .6 .8 1
X
B. Nonlinear E[Y
[ 0i| Xi]
1 1.5
e
0
0 .2 .4 .6 .8 1
X
B. Nonlinear E[Y
[ 0i| Xi]
1.51
Outcome
.5
O
0
0 .2 .4 .6 .8 1
X
Suppose that in addition to the assignment mechanism above, potential outcomes can
be described by a linear, constant-effects model.
The key difference between this regression and others we’ve used to estimate
treatment effects is that Di , the regressor of interest, is not only correlated with
Xi , it is a deterministic function of Xi .
Suppose that E[Y0i |xi ] = f (Xi ) for some reasonably smooth function, f (Xi ) (as
in Panel B)
0 .2 .4 .6 .8 1
X
To reduce the likelihood of such mistakes, we can look only at data in a neighborhood
around the discontinuity, say the interval [x0 − ∆, x0 + ∆] for some small number ∆.
Then we have
so that
Moreover, the validity of this nonparametric estimation strategy does not turn on
the constant effects assumption, Y1i − Y0i = ρ
the estimand in (20) is the average causal effect, E[Y1i − Y0i |Xi = x0 ]
The problem is that working in a small neighborhood of the cutoff means that you
don’t have many data points.