Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2025 Aug 5;122(31):e2413441122.
doi: 10.1073/pnas.2413441122. Epub 2025 Jul 31.

Data-driven equation discovery reveals nonlinear reinforcement learning in humans

Affiliations

Data-driven equation discovery reveals nonlinear reinforcement learning in humans

Kyle J LaFollette et al. Proc Natl Acad Sci U S A. .

Abstract

Computational models of reinforcement learning (RL) have significantly contributed to our understanding of human behavior and decision-making. Traditional RL models, however, often adopt a linear approach to updating reward expectations, potentially oversimplifying the nuanced relationship between human behavior and rewards. To address these challenges and explore models of RL, we utilized a method of model discovery using equation discovery algorithms. This method, currently used mainly in physics and biology, attempts to capture data by proposing a differential equation from an array of suggested linear and nonlinear functions. Using this method, we were able to identify a model of RL which we termed the Quadratic Q-Weighted model. The model suggests that reward prediction errors obey nonlinear dynamics and exhibit negativity biases, resulting in an underweighting of reward when expectations are low, and an overweighting of the absence of reward when expectations are high. We tested the generalizability of our model by comparing it to classical models used in nine published studies. Our model surpassed traditional models in predictive accuracy across eight out of these nine published datasets, demonstrating not only its generalizability but also its potential to offer insights into the complexities of human learning. This work showcases the integration of a behavioral task with advanced computational methodologies as a potent strategy for uncovering the intricate patterns of human cognition, marking a significant step forward in the development of computational models that are both interpretable and broadly applicable.

Keywords: dynamical systems; machine learning; nonlinear modeling; reinforcement learning.

PubMed Disclaimer

Conflict of interest statement

Competing interests statement:The authors declare no competing interest.

References

    1. Guest O., Martin A. E., How computational modeling can force theory building in psychological science. Perspect. Psychol. Sci. 16, 789–802 (2021). - PubMed
    1. Sutton R. S., Barto A. G., Reinforcement Learning: An Introduction (MIT Press, 1998).
    1. Akam T., Walton M. E., What is dopamine doing in model-based reinforcement learning? Curr. Opin. Behav. Sci. 38, 74–82 (2021). - PMC - PubMed
    1. Doll B. B., Simon D. A., Daw N. D., The ubiquity of model-based reinforcement learning. Curr. Opin. Neurobiol. 22, 1075–1081 (2012). - PMC - PubMed
    1. Gershman S. J., Uchida N., Believing in dopamine. Nat. Rev. Neurosci. 20, 703–714 (2019). - PMC - PubMed

LinkOut - more resources