Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2019 Apr;42(2):276-290.
doi: 10.1007/s10865-018-9964-1. Epub 2018 Aug 25.

Can the artificial intelligence technique of reinforcement learning use continuously-monitored digital data to optimize treatment for weight loss?

Affiliations

Can the artificial intelligence technique of reinforcement learning use continuously-monitored digital data to optimize treatment for weight loss?

Evan M Forman et al. J Behav Med. 2019 Apr.

Abstract

Behavioral weight loss (WL) trials show that, on average, participants regain lost weight unless provided long-term, intensive-and thus costly-intervention. Optimization solutions have shown mixed success. The artificial intelligence principle of "reinforcement learning" (RL) offers a new and more sophisticated form of optimization in which the intensity of each individual's intervention is continuously adjusted depending on patterns of response. In this pilot, we evaluated the feasibility and acceptability of a RL-based WL intervention, and whether optimization would achieve equivalent benefit at a reduced cost compared to a non-optimized intensive intervention. Participants (n = 52) completed a 1-month, group-based in-person behavioral WL intervention and then (in Phase II) were randomly assigned to receive 3 months of twice-weekly remote interventions that were non-optimized (NO; 10-min phone calls) or optimized (a combination of phone calls, text exchanges, and automated messages selected by an algorithm). The Individually-Optimized (IO) and Group-Optimized (GO) algorithms selected interventions based on past performance of each intervention for each participant, and for each group member that fit into a fixed amount of time (e.g., 1 h), respectively. Results indicated that the system was feasible to deploy and acceptable to participants and coaches. As hypothesized, we were able to achieve equivalent Phase II weight losses (NO = 4.42%, IO = 4.56%, GO = 4.39%) at roughly one-third the cost (1.73 and 1.77 coaching hours/participant for IO and GO, versus 4.38 for NO), indicating strong promise for a RL system approach to weight loss and maintenance.

Keywords: Artificial intelligence; Behavioral treatment; Lifestyle modification; Obesity; Optimization; Weight loss.

PubMed Disclaimer

Conflict of interest statement

Compliance with ethical standards

Conflict of interest Evan M. Forman, Stephanie G. Kerrigan, Meghan L. Butryn, Adrienne S. Juarascio, Stephanie M. Manasse, Santiago Ontañón, Diane H. Dallal, Rebecca J. Crochiere and Danielle Moskow declare that they have no conflict of interest.

Figures

Fig. 1
Fig. 1
Reinforcement learning system which continuously monitors responses to interventions and repeatedly selects interventions, for every participant, based on the exploit strategy (use existing knowledge to select most effective intervention so far) or the explore strategy (select an alternate intervention to learn to allow for changing contexts)
Fig. 2
Fig. 2
Time costs and weight losses by treatment condition

Similar articles

Cited by

References

    1. Abbeel P, & Ng AY (2004). Apprenticeship learning via inverse reinforcement learning. Paper presented at the Proceedings of the twenty-first international conference on Machine learning.
    1. Anderson JW, Vichitbandra S, Qian W, & Kryscio RJ (1999). Long-term weight maintenance after an intensive weight-loss program. Journal of the American College of Nutrition, 18, 620–627. - PubMed
    1. Association of Behavioral and Cognitive Therapies. (2018). Find a CBT Therapist. Retrieved May 23, 2018, from http://www.findcbt.org.
    1. Auer P, Cesa-Bianchi N, & Fischer P (2002). Finite-time analysis of the multiarmed bandit problem. Machine Learning, 47, 235–256.
    1. Aydin ME, & Öztemel E (2000). Dynamic job-shop scheduling using reinforcement learning agents. Robotics and Autonomous Systems, 33, 169–178.

Publication types