Corpus ID: 211146323

Double/Debiased Machine Learning for Dynamic Treatment Effects

  title={Double/Debiased Machine Learning for Dynamic Treatment Effects},
  author={Greg Lewis and Vasilis Syrgkanis},
We consider the estimation of treatment effects in settings when multiple treatments are assigned over time and treatments can have a causal effect on future outcomes. We formulate the problem as a linear state space Markov process with a high dimensional state and propose an extension of the double/debiased machine learning framework to estimate the dynamic effects of treatments. Our method allows the use of arbitrary machine learning methods to control for the high dimensional state, subject… Expand

Figures from this paper

Estimating the Long-Term Effects of Novel Treatments
Policy makers typically face the problem of wanting to estimate the long-term effects of novel treatments, while only having historical data of older treatment options. We assume access to aExpand
Causal Inference and Machine Learning in Practice with EconML and CausalML: Industrial Use Cases at Microsoft, TripAdvisor, Uber
In recent years, both academic research and industry applications see an increased effort in using machine learning methods to measure granular causal effects and design optimal policies based onExpand
Distributed Double Machine Learning with a Serverless Architecture
A prototype Python implementation DoubleML-Serverless is provided for the estimation of double machine learning models with the serverless computing platform AWS Lambda and its utility is demonstrated with a case study analyzing estimation times and costs. Expand
Regularizing Double Machine Learning in Partially Linear Endogenous Models
We estimate the linear coefficient in a partially linear model with confounding variables. We rely on double machine learning (DML) and extend it with an additional regularization and selectionExpand
Evaluating (weighted) dynamic treatment effects by double machine learning.
We consider evaluating the causal effects of dynamic treatments, i.e. of multiple treatment sequences in various periods, based on double machine learning to control for observed, time-varyingExpand


Double/Debiased Machine Learning for Treatment and Structural Parameters
We revisit the classic semiparametric problem of inference on a low dimensional parameter θ_0 in the presence of high-dimensional nuisance parameters η_0. We depart from the classical setting byExpand
One type of semiparametric regression is b8X A u(Z), where b and u(Z) are an unknown slope coefficient vector and function. Estimates of b based on incorrect parametrization of u are generallyExpand
Off-Policy Evaluation via Adaptive Weighting with Data from Contextual Bandits
This paper improves the DR estimator by adaptively weighting observations to control its variance, and shows that a t-statistic based on the improved estimator is asymptotically normal under certain conditions, allowing us to form confidence intervals and test hypotheses. Expand
Double Reinforcement Learning for Efficient Off-Policy Evaluation in Markov Decision Processes
This work considers for the first time the semiparametric efficiency limits of OPE in Markov decision processes (MDPs), where actions, rewards, and states are memoryless, and develops a new estimator based on cross-fold estimation of $q-functions and marginalized density ratios, which is term double reinforcement learning (DRL). Expand
Efficiently Breaking the Curse of Horizon in Off-Policy Evaluation with Double Reinforcement Learning
A new estimator based on Double Reinforcement Learning that leverages this structure for OPE and simultaneously uses estimated stationary density ratios and $q-functions and remains efficient when both are estimated at slow, nonparametric rates and remains consistent when either is estimated consistently. Expand
High-Dimensional Statistics: A Non-Asymptotic Viewpoint
This book provides a self-contained introduction to the area of high-dimensional statistics, aimed at the first-year graduate level, and includes chapters that are focused on core methodology and theory - including tail bounds, concentration inequalities, uniform laws and empirical process, and random matrices. Expand
Learning When-to-Treat Policies
An “advantage doubly robust” estimator for learning such dynamic treatment rules using observational data under the assumption of sequential ignorability is developed, which is practical for policy optimization, and does not need any structural assumptions. Expand
Model selection for G-estimation of Dynamic Treatment Regimes.
Using a quadratic approximation approach inspired by iteratively reweighted least squares, a quasi-likelihood function is derived for G-estimation within the DTR framework, and it is shown how it can be used to form an information criterion for blip model selection. Expand
Orthogonal Statistical Learning
By focusing on excess risk rather than parameter estimation, this work can give guarantees under weaker assumptions than in previous works and accommodate the case where the target parameter belongs to a complex nonparametric class. Expand
Regularized Orthogonal Machine Learning for Nonlinear Semiparametric Models
This paper contributes to the literature on high-dimensional sparse M-estimation by allowing the loss function to depend on a functional nuisance parameter, which we estimate by modern machineExpand