Generalization Bounds and Representation Learning for Estimation of Potential Outcomes and Causal Effects
Journal article, 2022

Practitioners in diverse fields such as healthcare, economics and education are eager to apply machine learning to improve decision making. The cost and impracticality of performing experiments and a recent monumental increase in electronic record keeping has brought attention to the problem of evaluating decisions based on non-experimental observational data. This is the setting of this work. In particular, we study estimation of individual-level potential outcomes and causal effects—such as a single patient’s response to alternative medication—from recorded contexts, decisions and outcomes. We give generalization bounds on the error in estimated outcomes based on distributional distance measures between re-weighted samples of groups receiving different treatments. We provide conditions under which our bounds are tight and show how they relate to results for unsupervised domain adaptation. Led by our theoretical results, we devise algorithms which learn representations and weighting functions that minimize our bounds by regularizing the representation’s induced treatment group distance, and encourage sharing of information between treatment groups. Finally, an experimental evaluation on real and synthetic data shows the value of our proposed representation architecture and regularization scheme.

overlap

Causal effects

generalization theory

domain adaptation

Author

Fredrik Johansson

Chalmers, Computer Science and Engineering (Chalmers), Data Science and AI

U. Shalit

Technion – Israel Institute of Technology

Nathan Kallus

Cornell University

D. Sontag

Massachusetts Institute of Technology (MIT)

Journal of Machine Learning Research

1532-4435 (ISSN) 1533-7928 (eISSN)

Vol. 23 166 1-50

Subject Categories

Economics

Applied Psychology

Biomedical Laboratory Science/Technology

Probability Theory and Statistics

More information

Latest update

10/27/2023