Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2023 Aug:216:1047-1057.

Assessing the Impact of Context Inference Error and Partial Observability on RL Methods for Just-In-Time Adaptive Interventions

Affiliations

Assessing the Impact of Context Inference Error and Partial Observability on RL Methods for Just-In-Time Adaptive Interventions

Karine Karine et al. Proc Mach Learn Res. 2023 Aug.

Abstract

Just-in-Time Adaptive Interventions (JITAIs) are a class of personalized health interventions developed within the behavioral science community. JITAIs aim to provide the right type and amount of support by iteratively selecting a sequence of intervention options from a pre-defined set of components in response to each individual's time varying state. In this work, we explore the application of reinforcement learning methods to the problem of learning intervention option selection policies. We study the effect of context inference error and partial observability on the ability to learn effective policies. Our results show that the propagation of uncertainty from context inferences is critical to improving intervention efficacy as context uncertainty increases, while policy gradient algorithms can provide remarkable robustness to partially observed behavioral state information.

Keywords: Reinforcement learning; adaptive interventions; context inference; empirical evaluation; mobile health; partial observability.

PubMed Disclaimer

Figures

Figure 1:
Figure 1:
Context inference error as a function of σ.
Figure 2:
Figure 2:
(a) Effect of learning with most likely context and context probabilities for DQN. (b) Effect of learning with most likely context and context probabilities for REINFORCE. (c) Effect of learning with most likely contexts and partial observability for REINFORCE and DQN. (d) Effect of learning with context probabilities and partial observability for REINFORCE and DQN.
Figure 3:
Figure 3:
The top row of plots shows the distribution of actions selected by REINFORCE when given access to context probabilities. The bottom row of plots shows the distribution of actions selected by REINFORCE when given access only to the inferred most likely context.
Figure 4:
Figure 4:
Learning curves of DQN and REINFORCE.
Figure 5:
Figure 5:
The top row of plots shows the distribution of actions selected by DQN when given access to context probabilities. The bottom row of plots shows the distribution of actions selected by DQN when given access only to the inferred most likely context.
Figure 6:
Figure 6:
Performance as a function of the disengagement increment ϵd and decay parameters δd, for DQN (top row) and REINFORCE (bottom row).

Similar articles

References

    1. Battalio Samuel L, Conroy David E, Dempsey Walter, Liao Peng, Menictas Marianne, Murphy Susan, Nahum-Shani Inbal, Qian Tianchen, Kumar Santosh, and Spring. Bonnie Sense2stop: a micro-randomized trial using wearable sensors to optimize a just-in-time-adaptive stress management intervention for smoking relapse prevention. Contemporary Clinical Trials, 109:106534, 2021. - PMC - PubMed
    1. Bruin Tim de, Kober Jens, Tuyls Karl, and Babuška. Robert The importance of experience replay database composition in deep reinforcement learning. In Deep Reinforcement Learning Workshop, Advances in Neural Information Processing Systems, 2015.
    1. Ertin Emre, Stohs Nathan, Kumar Santosh, Raij Andrew, Al’Absi Mustafa, and Shah. Siddharth Autosense: unobtrusively wearable sensor suite for inferring the onset, causality, and consequences of stress in the field. In Proceedings of the 9th ACM conference on embedded networked sensor systems, pages 274–287, 2011.
    1. Gönül Suat, Namlı Tuncay, Coşar Ahmet, and Toroslu İsmail Hakkı. A reinforcement learning based algorithm for personalization of digital, just-in-time, adaptive interventions. Artificial Intelligence in Medicine, 115:102062, 2021. - PubMed
    1. Hardeman Wendy, Houghton Julie, Lane Kathleen, Jones Andy, and Naughton Felix. A systematic review of just-in-time adaptive interventions (jitais) to promote physical activity. International Journal of Behavioral Nutrition and Physical Activity, 16(1):1–21, 2019. - PMC - PubMed

LinkOut - more resources