Instrumental variable
Encyclopedia
In statistics
Statistics
Statistics is the study of the collection, organization, analysis, and interpretation of data. It deals with all aspects of this, including the planning of data collection in terms of the design of surveys and experiments....

, econometrics
Econometrics
Econometrics has been defined as "the application of mathematics and statistical methods to economic data" and described as the branch of economics "that aims to give empirical content to economic relations." More precisely, it is "the quantitative analysis of actual economic phenomena based on...

, epidemiology
Epidemiology
Epidemiology is the study of health-event, health-characteristic, or health-determinant patterns in a population. It is the cornerstone method of public health research, and helps inform policy decisions and evidence-based medicine by identifying risk factors for disease and targets for preventive...

 and related disciplines, the method of instrumental variables (IV) is used to estimate causal relationships when controlled experiments are not feasible.

Instrumental variable methods allow consistent
Consistent estimator
In statistics, a sequence of estimators for parameter θ0 is said to be consistent if this sequence converges in probability to θ0...

 estimation when the explanatory variables
Dependent and independent variables
The terms "dependent variable" and "independent variable" are used in similar but subtly different ways in mathematics and statistics as part of the standard terminology in those subjects...

 (covariates) are correlated
Correlation
In statistics, dependence refers to any statistical relationship between two random variables or two sets of data. Correlation refers to any of a broad class of statistical relationships involving dependence....

 with the error terms
Errors and residuals in statistics
In statistics and optimization, statistical errors and residuals are two closely related and easily confused measures of the deviation of a sample from its "theoretical value"...

 of a regression
Regression analysis
In statistics, regression analysis includes many techniques for modeling and analyzing several variables, when the focus is on the relationship between a dependent variable and one or more independent variables...

 relationship. Such correlation may occur when the dependent variable causes at least one of the covariate
Covariate
In statistics, a covariate is a variable that is possibly predictive of the outcome under study. A covariate may be of direct interest or it may be a confounding or interacting variable....

s ("reverse" causation), when there are relevant explanatory variables which are omitted from the model, or when the covariates are subject to measurement error. In this situation, ordinary linear regression
Linear regression
In statistics, linear regression is an approach to modeling the relationship between a scalar variable y and one or more explanatory variables denoted X. The case of one explanatory variable is called simple regression...

 generally produces biased and inconsistent estimates. However, if an instrument is available, consistent estimates may still be obtained. An instrument is a variable that does not itself belong in the explanatory equation and is correlated with the endogenous explanatory variables, conditional on the other covariates. In linear models, there are two main requirements for using an IV:
  • The instrument must be correlated with the endogenous explanatory variables, conditional on the other covariates.
  • The instrument cannot be correlated with the error term in the explanatory equation, that is, the instrument cannot suffer from the same problem as the original predicting variable.

Definitions

Formal definitions of instrumental variables, using counterfactuals and graphical criteria, are given by Pearl
Judea Pearl
Judea Pearl is a computer scientist and philosopher, best known for developing the probabilistic approach to artificial intelligence and the development of Bayesian networks ....

 (2000). Notions of causality in econometrics, and their relationship with instrumental variables and other methods, are discussed by Heckman
James Heckman
James Joseph Heckman is an American economist and Nobel laureate. He is the Henry Schultz Distinguished Service Professor of Economics at the University of Chicago, Professor of Science and Society at University College Dublin and a Senior Research Fellow at the American Bar Foundation.Heckman...

 (2008)..

The theory of instrumental variables was first derived by Philip G. Wright in his 1928 book The Tariff on Animal and Vegetable Oils.

Example

Informally, in attempting to estimate the causal effect of some variable x on another y, an instrument is a third variable z which affects y only through its effect on x. For example, suppose a researcher wishes to estimate the causal effect of smoking on general health (as in Leigh and Schembri 2004). Correlation between health and smoking does not imply that smoking causes poor health because other variables may affect both health and smoking, or because health may affect smoking in addition to smoking causing health problems. It is at best difficult and expensive to conduct controlled experiments on smoking status in the general population. The researcher may proceed to attempt to estimate the causal effect of smoking on health from observational data by using the tax rate on tobacco products as an instrument for smoking in a health regression. If tobacco taxes only affect health because they affect smoking (holding other variables in the model fixed), correlation between tobacco taxes and health is evidence that smoking causes changes in health. An estimate of the effect of smoking on health can be made by also making use of the correlation between taxes and smoking patterns.

Applications

IV methods are commonly used to estimate causal effects in contexts in which controlled experiments are not available. Credibility of the estimates hinges on the selection of suitable instruments. Good instruments are often created by policy changes, for example, the cancellation of federal student aid scholarship program may reveal the effects of aid on some students' outcomes. Other natural and quasi-natural experiments
Natural experiment
A natural experiment is an observational study in which the assignment of treatments to subjects has been haphazard: That is, the assignment of treatments has been made "by nature", but not by experimenters. Thus, a natural experiment is not a controlled experiment...

 of various types are commonly exploited, for example, Miguel, Satyanath, and Sergenti (2004) use weather shocks to identify the effect of changes in economic growth (i.e., declines) on civil conflict. Angrist
Joshua Angrist
Joshua David Angrist is an American economist. He is Professor of Economics at the Massachusetts Institute of Technology. Angrist, a labor economist, has taught at the university level since 1989...

 and Krueger (2001)
presents a survey of the history and uses of instrumental variable techniques.

Estimation

Suppose the data are generated by a process of the form


where i indexes observations, is the dependent variable, is a covariate, is an unobserved error term representing all causes of other than , and is an unobserved scalar parameter. The parameter is the causal effect on of a one unit change in , holding all other causes of constant. The econometric goal is to estimate . For simplicity's sake assume the draws of are uncorrelated and that they are drawn from distributions with the same variance
Variance
In probability theory and statistics, the variance is a measure of how far a set of numbers is spread out. It is one of several descriptors of a probability distribution, describing how far the numbers lie from the mean . In particular, the variance is one of the moments of a distribution...

, that is, that the errors are serially uncorrelated and homoskedastic.

Suppose also that a regression model of nominally the same form is proposed. Given a random sample of T observations from this process, the ordinary least squares
Least squares
The method of least squares is a standard approach to the approximate solution of overdetermined systems, i.e., sets of equations in which there are more equations than unknowns. "Least squares" means that the overall solution minimizes the sum of the squares of the errors made in solving every...

 estimator is


where x, y and denote column vectors of length T. When x and are uncorrelated
Correlation
In statistics, dependence refers to any statistical relationship between two random variables or two sets of data. Correlation refers to any of a broad class of statistical relationships involving dependence....

, under certain regularity conditions the second term has an expected value conditional on x of zero and converges to zero in the limit, so the estimator is unbiased and consistent. When x and the other unmeasured, causal variables collapsed into the term are correlated, however, the OLS estimator is generally biased and inconsistent for β. In this case, it is valid to use the estimates to predict values of y given values of x, but the estimate does not recover the causal effect of x on y.

An instrumental variable z is one that is correlated with the independent variable but not with the error
Errors and residuals in statistics
In statistics and optimization, statistical errors and residuals are two closely related and easily confused measures of the deviation of a sample from its "theoretical value"...

 term. Using the method of moments, take expectations conditional on z to find


The second term on the right-hand side is zero by assumption. Solve for and write the resulting expression in terms of sample moments,


When z and are uncorrelated, the final term, under certain regularity conditions, approaches zero in the limit, providing a consistent estimator. Put another way, the causal effect of x on y can be consistently estimated from these data even though x is not randomly assigned through experimental methods.

The approach generalizes to a model with multiple explanatory variables. Suppose X is the T × K matrix of explanatory variables resulting from T observations on K variables. Let Z be a T × K matrix of instruments. Then it can be shown that the estimator


is consistent under a multivariate generalization of the conditions discussed above. If there are more instruments than there are covariates in the equation of interest so that Z is a T × M matrix with M > K, the generalized method of moments
Generalized method of moments
In econometrics, generalized method of moments is a generic method for estimating parameters in statistical models. Usually it is applied in the context of semiparametric models, where the parameter of interest is finite-dimensional, whereas the full shape of the distribution function of the data...

 can be used and the resulting IV estimator is


where . The second expression collapses to the first when the number of instruments is equal to the number of covariates in the equation of interest.

Interpretation as two-stage least squares

One computational method which can be used to calculate IV estimates is two-stage least-squares (2SLS). In the first stage, each endogenous covariate in the equation of interest is regressed on all of the exogenous variables in the model, including both exogenous covariates in the equation of interest and the excluded instruments. The predicted values from these regressions are obtained.

Stage 1: Regress each column of X on Z, ()


and save the predicted values:


In the second stage, the regression of interest is estimated as usual, except that in this stage each endogenous covariate is replaced with the predicted values from its first stage model from the first stage.

Stage 2: Regress Y on the predicted values from the first stage:


The resulting estimator of is numerically identical to the expression displayed above. A small correction must be made to the sum-of-squared residuals in the second-stage fitted model in order that the covariance matrix of is calculated correctly.

Identification

In the instrumental variable regression, if we have multiple endogenous regressors and multiple instruments the coefficients on the endogenous regressors are said to be:

Exactly identified if m = k.

Overidentified if m > k.

Underidentified if m < k.

The parameters are not identified if there are fewer instruments than there are covariates or, equivalently, if there are fewer excluded instruments than there are endogenous covariates in the equation of interest.

Non-parametric analysis

When the form of the structural equations is unknown, an instrumental variable can still be defined through the equations:
where and are two arbitrary functions and is
independent of . Unlike linear models, however, measurements
of and do not allow for the identification of the
average causal effect of on , denoted ACE
Balke and Pearl [1997] derived tight bounds on
ACE and showed that these can provide valuable information
on the sign and size of ACE.

In linear analysis, there is no test to falsify
the assumption the is instrumental relative to
the pair . This is not the case when is
discrete. Pearl (2000) has shown that, for all
and , the following constraint, called "Instrumental
Inequality" must hold whenever satisfies the two equations above:

On the interpretation of IV estimates

The exposition above assumes that the causal effect of interest does not vary across observations, that is, that is a constant. Generally, different subjects will respond differently to changes in the "treatment" x. When this possibility is recognized, the average effect in the population of a change in x on y may differ from the effect in a given subpopulation. For example, the average effect of a job training program may substantially differ across the group of people who actually receive the training and the group which chooses not to receive training. For these reasons, IV methods invoke implicit assumptions on behavioral response, or more generally assumptions over the correlation between the response to treatment and propensity to receive treatment.

The standard IV estimator can recover local average treatment effects (LATE) rather than average treatment effects
Average treatment effects
The average treatment effect is a measure used to compare treatments in randomized experiments, evaluation of policy interventions, and medical trials. The ATE measures the average causal difference in outcomes under the treatment and under the control...

 (ATE). Imbens and Angrist (1994) demonstrate that the linear IV estimate can be interpreted under weak conditions as a weighted average of local average treatment effects, where the weights depend on the elasticity of the endogenous regressor to changes in the instrumental variables. Roughly, that means that the effect of a variable is only revealed for the subpopulations affected by the observed changes in the instruments, and that subpopulations which respond most to changes in the instruments will have the largest effects on the magnitude of the IV estimate.

For example, if a researcher uses presence of a land-grant college as an instrument for college education in an earnings regression, she identifies the effect of college on earnings in the subpopulation which would obtain a college degree if a college is present but which would not obtain a degree if a college is not present. This empirical approach does not, without further assumptions, tell the researcher anything about the effect of college among people who would either always or never get a college degree regardless of whether a local college exists.

Potential problems

Instrumental variables estimates are generally inconsistent if the instruments are correlated with the error term in the equation of interest. Another problem is caused by the selection of "weak" instruments, instruments that are poor predictors of the endogenous question predictor in the first-stage equation. In this case, the prediction of the question predictor by the instrument will be poor and the predicted values will have very little variation. Consequently, they are unlikely to have much success in predicting the ultimate outcome when they are used to replace the question predictor in the second-stage equation.

In the context of the smoking and health example discussed above, tobacco taxes are weak instruments for smoking if smoking status is largely unresponsive to changes in taxes. If higher taxes do not induce people to quit smoking (or not start smoking), then variation in tax rates tells us nothing about the effect of smoking on health. If taxes affect health through channels other than through their effect on smoking, then the instruments are invalid and the instrumental variables approach may yield misleading results. For example, places and times with relatively health-conscious populations may both implement high tobacco taxes and exhibit better health even holding smoking rates constant, so we would observe a correlation between health and tobacco taxes even if it were the case that smoking has no effect on health. In this case, we would be mistaken to infer a causal effect of smoking on health from the observed correlation between tobacco taxes and health.

Sampling properties and hypothesis testing

When the covariates are exogenous, the small-sample properties of the OLS estimator can be derived in a straightforward manner by calculating moments of the estimator conditional on X. When some of the covariates are endogenous so that instrumental variables estimation is implemented, simple expressions for the moments of the estimator cannot be so obtained. Generally, instrumental variables estimators only have desirable asymptotic, not finite sample, properties, and inference is based on asymptotic approximations to the sampling distribution of the estimator. Even when the instruments are uncorrelated with the error in the equation of interest and when the instruments are not weak, the finite sample properties of the instrumental variables estimator may be poor. For example, exactly identified models produce finite sample estimators with no moments, so the estimator can be said to be neither biased nor unbiased, the nominal size of test statistics may be substantially distorted, and the estimates may commonly be far away from the true value of the parameter (Nelson and Startz 1990).

Testing instrument strength and overidentifying restrictions

The strength of the instruments can be directly assessed because both the endogenous covariates and the instruments are observable (Stock, Wright, and Yogo 2002). A common rule of thumb for models with one endogenous regressor is: the F-statistic against the null that the excluded instruments are irrelevant in the first-stage regression should be larger than 10.

The assumption that the instruments are not correlated with the error term in the equation of interest is not testable in exactly identified models. If the model is overidentified, there is information available which may be used to test this assumption. The most common test of these overidentifying restrictions, called the Sargan test
Sargan test
The Sargan test is a statistical test used to check for over-identifying restrictions in a statistical model. The Sargan test is based on the observation that the residuals should be uncorrelated with the set of exogenous variables if the instruments are truly exogenous...

, is based on the observation that the residuals should be uncorrelated with the set of exogenous variables if the instruments are truly exogenous. The Sargan test statistic can be calculated as (the number of observations multiplied by the coefficient of determination
Coefficient of determination
In statistics, the coefficient of determination R2 is used in the context of statistical models whose main purpose is the prediction of future outcomes on the basis of other related information. It is the proportion of variability in a data set that is accounted for by the statistical model...

) from the OLS regression of the residuals onto the set of exogenous variables. This statistic will be asymptotically chi-squared with m − k degrees of freedom under the null that the error term is uncorrelated with the instruments.

External links

  • Layman's explanation of instrumental variables.
  • Chapter from Daniel McFadden
    Daniel McFadden
    Daniel Little McFadden is an econometrician who shared the 2000 Nobel Memorial Prize in Economic Sciences with James Heckman ; McFadden's share of the prize was "for his development of theory and methods for analyzing discrete choice". He was the E. Morris Cox Professor of Economics at the...

    's textbook
The source of this article is wikipedia, the free encyclopedia.  The text of this article is licensed under the GFDL.
 
x
OK