Long Story Short: Omitted Variable Bias in Causal Machine Learning / Victor Chernozhukov, Carlos Cinelli, Whitney Newey, Amit Sharma, Vasilis Syrgkanis.
Material type:![Text](/opac-tmpl/lib/famfamfam/BK.png)
- Semiparametric and Nonparametric Methods: General
- Semiparametric and Nonparametric Methods: General
- Cross-Sectional Models • Spatial Models • Treatment Effect Models • Quantile Regressions
- Cross-Sectional Models • Spatial Models • Treatment Effect Models • Quantile Regressions
- Cross-Sectional Models • Spatial Models • Treatment Effect Models • Quantile Regressions • Social Interaction Models
- Cross-Sectional Models • Spatial Models • Treatment Effect Models • Quantile Regressions • Social Interaction Models
- C14
- C21
- C31
- Hardcopy version available to institutional subscribers
Item type | Home library | Collection | Call number | Status | Date due | Barcode | Item holds | |
---|---|---|---|---|---|---|---|---|
Working Paper | Biblioteca Digital | Colección NBER | nber w30302 (Browse shelf(Opens below)) | Not For Loan |
July 2022.
We derive general, yet simple, sharp bounds on the size of the omitted variable bias for a broad class of causal parameters that can be identified as linear functionals of the conditional expectation function of the outcome. Such functionals encompass many of the traditional targets of investigation in causal inference studies, such as, for example, (weighted) average of potential outcomes, average treatment effects (including subgroup effects, such as the effect on the treated), (weighted) average derivatives, and policy effects from shifts in covariate distribution -- all for general, nonparametric causal models. Our construction relies on the Riesz-Frechet representation of the target functional. Specifically, we show how the bound on the bias depends only on the additional variation that the latent variables create both in the outcome and in the Riesz representer for the parameter of interest. Moreover, in many important cases (e.g, average treatment effects and avearage derivatives) the bound is shown to depend on easily interpretable quantities that measure the explanatory power of the omitted variables. Therefore, simple plausibility judgments on the maximum explanatory power of omitted variables (in explaining treatment and outcome variation) are sufficient to place overall bounds on the size of the bias. Furthermore, we use debiased machine learning to provide flexible and efficient statistical inference on learnable components of the bounds. Finally, empirical examples demonstrate the usefulness of the approach.
Hardcopy version available to institutional subscribers
System requirements: Adobe [Acrobat] Reader required for PDF files.
Mode of access: World Wide Web.
Print version record
There are no comments on this title.