Reinforcement learning and Bayesian inference provide complementary models for the unique advantage of adolescents in stochastic reversal
- PMID: 35537273
- PMCID: PMC9108470
- DOI: 10.1016/j.dcn.2022.101106
Reinforcement learning and Bayesian inference provide complementary models for the unique advantage of adolescents in stochastic reversal
Abstract
During adolescence, youth venture out, explore the wider world, and are challenged to learn how to navigate novel and uncertain environments. We investigated how performance changes across adolescent development in a stochastic, volatile reversal-learning task that uniquely taxes the balance of persistence and flexibility. In a sample of 291 participants aged 8-30, we found that in the mid-teen years, adolescents outperformed both younger and older participants. We developed two independent cognitive models, based on Reinforcement learning (RL) and Bayesian inference (BI). The RL parameter for learning from negative outcomes and the BI parameters specifying participants' mental models were closest to optimal in mid-teen adolescents, suggesting a central role in adolescent cognitive processing. By contrast, persistence and noise parameters improved monotonically with age. We distilled the insights of RL and BI using principal component analysis and found that three shared components interacted to form the adolescent performance peak: adult-like behavioral quality, child-like time scales, and developmentally-unique processing of positive feedback. This research highlights adolescence as a neurodevelopmental window that can create performance advantages in volatile and uncertain environments. It also shows how detailed insights can be gleaned by using cognitive models in new ways.
Keywords: Adolescence; Bayesian inference; Computational modeling; Development; Non-linear changes; Reinforcement learning; Volatility.
Copyright © 2022 The Authors. Published by Elsevier Ltd.. All rights reserved.
Conflict of interest statement
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Figures






Similar articles
-
Intact Reinforcement Learning But Impaired Attentional Control During Multidimensional Probabilistic Learning in Older Adults.J Neurosci. 2020 Jan 29;40(5):1084-1096. doi: 10.1523/JNEUROSCI.0254-19.2019. Epub 2019 Dec 11. J Neurosci. 2020. PMID: 31826943 Free PMC article.
-
Computational evidence for hierarchically structured reinforcement learning in humans.Proc Natl Acad Sci U S A. 2020 Nov 24;117(47):29381-29389. doi: 10.1073/pnas.1912330117. Proc Natl Acad Sci U S A. 2020. PMID: 33229518 Free PMC article.
-
Reinforcement Learning during Adolescence in Rats.J Neurosci. 2020 Jul 22;40(30):5857-5870. doi: 10.1523/JNEUROSCI.0910-20.2020. Epub 2020 Jun 29. J Neurosci. 2020. PMID: 32601244 Free PMC article.
-
Computational approaches to modeling gambling behaviour: Opportunities for understanding disordered gambling.Neurosci Biobehav Rev. 2023 Apr;147:105083. doi: 10.1016/j.neubiorev.2023.105083. Epub 2023 Feb 8. Neurosci Biobehav Rev. 2023. PMID: 36758827 Review.
-
Social learning across adolescence: A Bayesian neurocognitive perspective.Dev Cogn Neurosci. 2022 Dec;58:101151. doi: 10.1016/j.dcn.2022.101151. Epub 2022 Sep 16. Dev Cogn Neurosci. 2022. PMID: 36183664 Free PMC article. Review.
Cited by
-
The rational use of causal inference to guide reinforcement learning strengthens with age.NPJ Sci Learn. 2020 Oct 27;5:16. doi: 10.1038/s41539-020-00075-3. eCollection 2020. NPJ Sci Learn. 2020. PMID: 33133638 Free PMC article.
-
Revisiting the role of computational neuroimaging in the era of integrative neuroscience.Neuropsychopharmacology. 2024 Nov;50(1):103-113. doi: 10.1038/s41386-024-01946-8. Epub 2024 Sep 6. Neuropsychopharmacology. 2024. PMID: 39242921 Free PMC article. Review.
-
Nucleus accumbens dopamine release reflects Bayesian inference during instrumental learning.bioRxiv [Preprint]. 2024 Sep 13:2023.11.10.566306. doi: 10.1101/2023.11.10.566306. bioRxiv. 2024. Update in: PLoS Comput Biol. 2025 Jul 2;21(7):e1013226. doi: 10.1371/journal.pcbi.1013226. PMID: 38014354 Free PMC article. Updated. Preprint.
-
Action initiation and punishment learning differ from childhood to adolescence while reward learning remains stable.Nat Commun. 2023 Sep 14;14(1):5689. doi: 10.1038/s41467-023-41124-w. Nat Commun. 2023. PMID: 37709750 Free PMC article.
-
Nucleus accumbens dopamine release reflects Bayesian inference during instrumental learning.PLoS Comput Biol. 2025 Jul 2;21(7):e1013226. doi: 10.1371/journal.pcbi.1013226. eCollection 2025 Jul. PLoS Comput Biol. 2025. PMID: 40601769 Free PMC article.
References
-
- Bates D., Mächler M., Bolker B., Walker S. Fitting linear mixed-effects models using lme4. J. Stat. Softw. 2015;67(1):1–48. doi: 10.18637/jss.v067.i01. - DOI
-
- Bernardo J.M., Smith A.F.M. John Wiley & Sons; 2009. Bayesian Theory. Google-Books-ID: 11nSgIcd7xQC.