scispace - formally typeset
Open AccessJournal ArticleDOI

Recursive partitioning for heterogeneous causal effects

Susan Athey, +1 more
- 05 Jul 2016 - 
- Vol. 113, Iss: 27, pp 7353-7360
Reads0
Chats0
TLDR
This paper provides a data-driven approach to partition the data into subpopulations that differ in the magnitude of their treatment effects, and proposes an “honest” approach to estimation, whereby one sample is used to construct the partition and another to estimate treatment effects for each subpopulation.
Abstract
In this paper we propose methods for estimating heterogeneity in causal effects in experimental and observational studies and for conducting hypothesis tests about the magnitude of differences in treatment effects across subsets of the population. We provide a data-driven approach to partition the data into subpopulations that differ in the magnitude of their treatment effects. The approach enables the construction of valid confidence intervals for treatment effects, even with many covariates relative to the sample size, and without “sparsity” assumptions. We propose an “honest” approach to estimation, whereby one sample is used to construct the partition and another to estimate treatment effects for each subpopulation. Our approach builds on regression tree methods, modified to optimize for goodness of fit in treatment effects and to account for honest estimation. Our model selection criterion anticipates that bias will be eliminated by honest estimation and also accounts for the effect of making additional splits on the variance of treatment effect estimates within each subpopulation. We address the challenge that the “ground truth” for a causal effect is not observed for any individual unit, so that standard approaches to cross-validation must be modified. Through a simulation study, we show that for our preferred method honest estimation results in nominal coverage for 90% confidence intervals, whereas coverage ranges between 74% and 84% for nonhonest approaches. Honest estimation requires estimating the model with a smaller sample size; the cost in terms of mean squared error of treatment effects for our preferred method ranges between 7–22%.

read more

Citations
More filters
Journal ArticleDOI

Heterogeneity in the Effect of Environmental Protection Expenditure in China: Causal Inference from Machine Learning

TL;DR: Li et al. as mentioned in this paper used new machine learning to estimate heterogeneous treatment effect of the EPE shock on urban air pollution from 216 cities in China, from 2011 to 2018.
Journal ArticleDOI

Targeting resources efficiently and justifiably by combining causal machine learning and theory

TL;DR: In this paper , a case study identifies the right individuals to incentivize for increasing their physical activity to maximize the population's health benefits due to reduced diabetes and heart disease prevalence, using causal machine learning to estimate heterogeneous treatment effects from big data.
Posted Content

Estimation of Discrete Choice Models: A Machine Learning Approach

TL;DR: This paper proposes a new method of estimation for discrete choice demand models when individual level data are available and uses nonparametric approximation (followed by) moment estimation, hence the name---NAME.
Dissertation

Estimating Causal Effects in Pragmatic Settings With Imperfect Information

David Cheng
Abstract: Precision medicine seeks to identify the optimal treatment for each individual based on his or her unique features. This invariably involves some form of estimation of causal effects for different patient subgroups to determine the treatment that leads to superior outcomes. Implementing methods to estimate causal effects in modern large and rich data sources such as electronic medical records (EMR), however, still faces challenges as information on patients is imperfectly captured in the observed data. In this work, we propose approaches to address some of the primary issues encountered in estimating causal effects in these pragmatic settings. In Chapter 1, we consider estimating average treatment effects (ATE) in observational data where the number of covariates is not small relative to the sample size. We develop a double-index propensity score (DiPS) obtained by smoothing treatment over linear predictors for the covariates from initial working parametric propensity score (PS) and outcome models fit with regularization. We show that an inverse probability weighting (IPW) estimator based on DiPS maintains the doubly-robustness and local semiparametric efficiency properties of the usual doubly-robust estimator and achieves further gains in robustness and efficiency under model misspecification. Simulations demonstrate the benefit of the approach in finite samples, and the method is illustrated by applications estimating the effects of statins on colorectal cancer risk and smoking on C-reactive protein. In Chapter 2, we extend the work from Chapter 1 to allow for incorporation of a large set of unlabeled data. This arises in EMR data when chart review is performed to ascertain gold-standard outcomes in case outcomes of interest are not directly observed. We frame the problem in a semi-supervised learning setting, where a small set of observations
Journal ArticleDOI

The Effect of COVID-19 Vaccinations on Self-Reported Depression and Anxiety During February 2021

TL;DR: The average effect of COVID-19 vaccinations on self-reported feelings of depression and anxiety, isolation, and worries about health among vaccine-accepting respondents in February 2021, and find 3.7, 3.3, and 4.3 percentage point reductions in the probability of each outcome as discussed by the authors .
References
More filters
Journal ArticleDOI

Random Forests

TL;DR: Internal estimates monitor error, strength, and correlation and these are used to show the response to increasing the number of features used in the forest, and are also applicable to regression.
Journal ArticleDOI

Regression Shrinkage and Selection via the Lasso

TL;DR: A new method for estimation in linear models called the lasso, which minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant, is proposed.
Book

The Nature of Statistical Learning Theory

TL;DR: Setting of the learning problem consistency of learning processes bounds on the rate of convergence ofLearning processes controlling the generalization ability of learning process constructing learning algorithms what is important in learning theory?

Statistical learning theory

TL;DR: Presenting a method for determining the necessary and sufficient conditions for consistency of learning process, the author covers function estimates from small data pools, applying these estimations to real-life problems, and much more.
Journal ArticleDOI

The central role of the propensity score in observational studies for causal effects

Paul R. Rosenbaum, +1 more
- 01 Apr 1983 - 
TL;DR: The authors discusses the central role of propensity scores and balancing scores in the analysis of observational studies and shows that adjustment for the scalar propensity score is sufficient to remove bias due to all observed covariates.