Uncertainty and exploration in a restless bandit problem
- PMID: 25899069
- DOI: 10.1111/tops.12145
Uncertainty and exploration in a restless bandit problem
Abstract
Decision making in noisy and changing environments requires a fine balance between exploiting knowledge about good courses of action and exploring the environment in order to improve upon this knowledge. We present an experiment on a restless bandit task in which participants made repeated choices between options for which the average rewards changed over time. Comparing a number of computational models of participants' behavior in this task, we find evidence that a substantial number of them balanced exploration and exploitation by considering the probability that an option offers the maximum reward out of all the available options.
Keywords: Dynamic decision making; Exploration-exploitation trade-off; Restless multi-armed bandit task; Uncertainty; Volatility.
Copyright © 2015 Cognitive Science Society, Inc.
MeSH terms
LinkOut - more resources
Full Text Sources
