» Articles » PMID: 36396808

An Interpretable RL Framework for Pre-deployment Modeling in ICU Hypotension Management

Overview
Journal NPJ Digit Med
Date 2022 Nov 17
PMID 36396808
Authors
Affiliations
Soon will be listed here.
Abstract

Computational methods from reinforcement learning have shown promise in inferring treatment strategies for hypotension management and other clinical decision-making challenges. Unfortunately, the resulting models are often difficult for clinicians to interpret, making clinical inspection and validation of these computationally derived strategies challenging in advance of deployment. In this work, we develop a general framework for identifying succinct sets of clinical contexts in which clinicians make very different treatment choices, tracing the effects of those choices, and inferring a set of recommendations for those specific contexts. By focusing on these few key decision points, our framework produces succinct, interpretable treatment strategies that can each be easily visualized and verified by clinical experts. This interrogation process allows clinicians to leverage the model's use of historical data in tandem with their own expertise to determine which recommendations are worth investigating further e.g. at the bedside. We demonstrate the value of this approach via application to hypotension management in the ICU, an area with critical implications for patient outcomes that lacks data-driven individualized treatment strategies; that said, our framework has broad implications on how to use computational methods to assist with decision-making challenges on a wide range of clinical domains.

Citing Articles

Safety of human-AI cooperative decision-making within intensive care: A physical simulation study.

Festor P, Nagendran M, Gordon A, Faisal A, Komorowski M PLOS Digit Health. 2025; 4(2):e0000726.

PMID: 39992918 PMC: 11849858. DOI: 10.1371/journal.pdig.0000726.


Personalized decision making for coronary artery disease treatment using offline reinforcement learning.

Ghasemi P, Greenberg M, Southern D, Li B, White J, Lee J NPJ Digit Med. 2025; 8(1):99.

PMID: 39948243 PMC: 11825836. DOI: 10.1038/s41746-025-01498-1.


Reinforcement learning model for optimizing dexmedetomidine dosing to prevent delirium in critically ill patients.

Lee H, Chung S, Hyeon D, Yang H, Lee H, Ryu H NPJ Digit Med. 2024; 7(1):325.

PMID: 39557970 PMC: 11574043. DOI: 10.1038/s41746-024-01335-x.


Development and validation of a reinforcement learning model for ventilation control during emergence from general anesthesia.

Lee H, Yoon H, Kim J, Park J, Koo C, Won D NPJ Digit Med. 2023; 6(1):145.

PMID: 37580410 PMC: 10425339. DOI: 10.1038/s41746-023-00893-w.

References
1.
Oh S, Lee S, Noh J, Mo J . Optimal treatment recommendations for diabetes patients using the Markov decision process along with the South Korean electronic health records. Sci Rep. 2021; 11(1):6920. PMC: 7994640. DOI: 10.1038/s41598-021-86419-4. View

2.
Lage I, Lifschitz D, Doshi-Velez F, Amir O . Exploring Computational User Models for Agent Policy Summarization. IJCAI (U S). 2021; 28:1401-1407. PMC: 7901848. View

3.
Komorowski M, Celi L, Badawi O, Gordon A, Faisal A . The Artificial Intelligence Clinician learns optimal treatment strategies for sepsis in intensive care. Nat Med. 2018; 24(11):1716-1720. DOI: 10.1038/s41591-018-0213-5. View

4.
Kox M, Pickkers P . "Less is more" in critically ill patients: not too intensive. JAMA Intern Med. 2013; 173(14):1369-72. DOI: 10.1001/jamainternmed.2013.6702. View

5.
Kourou K, Exarchos T, Exarchos K, Karamouzis M, Fotiadis D . Machine learning applications in cancer prognosis and prediction. Comput Struct Biotechnol J. 2015; 13:8-17. PMC: 4348437. DOI: 10.1016/j.csbj.2014.11.005. View