Y Cui, H Pu, X Shi, W Miao… - Journal of the American …, 2024 - Taylor & Francis
Skepticism about the assumption of no unmeasured confounding, also known as exchangeability, is often warranted in making causal inferences from observational data; …
Abstract We study Reinforcement Learning for partially observable systems using function approximation. We propose a new PO-bilinear framework, that is general enough to include …
We study the estimation of causal parameters when not all confounders are observed and instead negative controls are available. Recent work has shown how these can enable …
C Shi, M Uehara, J Huang… - … Conference on Machine …, 2022 - proceedings.mlr.press
We consider off-policy evaluation (OPE) in Partially Observable Markov Decision Processes (POMDPs), where the evaluation policy depends only on observable variables and the …
J Kaddour, Y Zhu, Q Liu… - Advances in Neural …, 2021 - proceedings.neurips.cc
We address the estimation of conditional average treatment effects (CATEs) for structured treatments (eg, graphs, images, texts). Given a weak condition on the effect, we propose the …
A Ying, W Miao, X Shi… - Journal of the Royal …, 2023 - academic.oup.com
A standard assumption for causal inference about the joint effects of time-varying treatment is that one has measured sufficient covariates to ensure that within covariate strata, subjects …
We study offline reinforcement learning (RL) in partially observable Markov decision processes. In particular, we aim to learn an optimal policy from a dataset collected by a …
We study the identification and estimation of long-term treatment effects when both experimental and observational data are available. Since the long-term outcome is …
Proxy causal learning (PCL) is a method for estimating the causal effect of treatments on outcomes in the presence of unobserved confounding, using proxies (structured side …