Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 Nov 18;5(1):173.
doi: 10.1038/s41746-022-00708-4.

An interpretable RL framework for pre-deployment modeling in ICU hypotension management

Affiliations

An interpretable RL framework for pre-deployment modeling in ICU hypotension management

Kristine Zhang et al. NPJ Digit Med. .

Abstract

Computational methods from reinforcement learning have shown promise in inferring treatment strategies for hypotension management and other clinical decision-making challenges. Unfortunately, the resulting models are often difficult for clinicians to interpret, making clinical inspection and validation of these computationally derived strategies challenging in advance of deployment. In this work, we develop a general framework for identifying succinct sets of clinical contexts in which clinicians make very different treatment choices, tracing the effects of those choices, and inferring a set of recommendations for those specific contexts. By focusing on these few key decision points, our framework produces succinct, interpretable treatment strategies that can each be easily visualized and verified by clinical experts. This interrogation process allows clinicians to leverage the model's use of historical data in tandem with their own expertise to determine which recommendations are worth investigating further e.g. at the bedside. We demonstrate the value of this approach via application to hypotension management in the ICU, an area with critical implications for patient outcomes that lacks data-driven individualized treatment strategies; that said, our framework has broad implications on how to use computational methods to assist with decision-making challenges on a wide range of clinical domains.

PubMed Disclaimer

Conflict of interest statement

The authors declare competing interests: F.D.V. consults for Davita Kidney Care. K.Z., H.W., J.D., B.C., L.A.C., and R.K. declare no competing interests.

Figures

Fig. 1
Fig. 1. The decision point pipeline.
a We optimize a similarity metric to reflect physicians' perceptions of patient similarity. b We identify decision regions (DR), or areas where similar patients frequently receive different treatments. c We summarize patient trajectories in terms of decision regions. d We use this Markov Decision Process to learn an optimal treatment policy for each decision region.
Fig. 2
Fig. 2. Probability assigned to each action under policies.
It is obtained using three different reward functions, compared to the current practices from clinicians.
Fig. 3
Fig. 3. The feature means for points in each decision region.
Means are sorted by descending order of Mean Arterial Pressure (MAP) value.
Fig. 4
Fig. 4. Change in patient state after different treatments.
a The probability of patients moving from one decision region to other decision regions when given different treatments. This example figure shows how patients transition out of decision region 9. b The expected average feature value for patients in each decision region after different treatments were given. This example figure reflects MAP level change.
Fig. 5
Fig. 5. Examples of patient trajectories.
For three specific patient cases where the recommended action differs from the clinician actions, we track the changes over time in their clinical metrics.
Fig. 6
Fig. 6. Feature importance of clinical variables.
Bar charts showing feature importance, derived from random forest classifiers, in descending order.

References

    1. Komorowski M, Celi LA, Badawi O, Gordon AC, Faisal AA. The Artificial Intelligence Clinician learns optimal treatment strategies for sepsis in intensive care. Nat. Med. 2018;24:1716–1720. doi: 10.1038/s41591-018-0213-5. - DOI - PubMed
    1. Yu, C., Liu, J., Nemati, S. & Yin, G. Reinforcement learning in healthcare: a survey. ACM Comput. Surv. 55, 5 (2021).
    1. Kourou K, Exarchos TP, Exarchos KP, Karamouzis MV, Fotiadis DI. Machine learning applications in cancer prognosis and prediction. Comput. Struct. Biotechnol. J. 2015;13:8–17. doi: 10.1016/j.csbj.2014.11.005. - DOI - PMC - PubMed
    1. Tseng, H.-H. et al. Deep reinforcement learning for automated radiation adaptation in lung cancer. Med. Phys.44, 6690–6705 (2017). - PMC - PubMed
    1. Oh, S.-H., Lee, S. J., Noh, J. & Mo, J. Optimal treatment recommendations for diabetes patients usingthe Markov decision process along with the South Korean electronic health records. Sci. Rep.11, 6920 (2021). - PMC - PubMed