| Literature DB >> 29873099 |
Ying Liu1, Yuanjia Wang2, Michael R Kosorok3, Yingqi Zhao4, Donglin Zeng3.
Abstract
Dynamic treatment regimens (DTRs) are sequential treatment decisions tailored by patient's evolving features and intermediate outcomes at each treatment stage. Patient heterogeneity and the complexity and chronicity of many diseases call for learning optimal DTRs that can best tailor treatment according to each individual's time-varying characteristics (eg, intermediate response over time). In this paper, we propose a robust and efficient approach referred to as Augmented Outcome-weighted Learning (AOL) to identify optimal DTRs from sequential multiple assignment randomized trials. We improve previously proposed outcome-weighted learning to allow for negative weights. Furthermore, to reduce the variability of weights for numeric stability and improve estimation accuracy, in AOL, we propose a robust augmentation to the weights by making use of predicted pseudooutcomes from regression models for Q-functions. We show that AOL still yields Fisher-consistent DTRs even if the regression models are misspecified and that an appropriate choice of the augmentation guarantees smaller stochastic errors in value function estimation for AOL than the previous outcome-weighted learning. Finally, we establish the convergence rates for AOL. The comparative advantage of AOL over existing methods is demonstrated through extensive simulation studies and an application to a sequential multiple assignment randomized trial for major depressive disorder.Entities:
Keywords: Q-learning; SMARTs; adaptive intervention; individualized treatment rule; machine learning; outcome-weighted learning; personalized medicine
Mesh:
Year: 2018 PMID: 29873099 PMCID: PMC6191367 DOI: 10.1002/sim.7844
Source DB: PubMed Journal: Stat Med ISSN: 0277-6715 Impact factor: 2.373