Mapping value based planning and extensively trained choice in the human brain
- PMID: 22406551
- PMCID: PMC3378641
- DOI: 10.1038/nn.3068
Mapping value based planning and extensively trained choice in the human brain
Abstract
Investigations of the underlying mechanisms of choice in humans have focused on learning from prediction errors, leaving the computational structure of value based planning comparatively underexplored. Using behavioral and neuroimaging analyses of a minimax decision task, we found that the computational processes underlying forward planning are expressed in the anterior caudate nucleus as values of individual branching steps in a decision tree. In contrast, values represented in the putamen pertain solely to values learned during extensive training. During actual choice, both striatal areas showed a functional coupling to ventromedial prefrontal cortex, consistent with this region acting as a value comparator. Our findings point toward an architecture of choice in which segregated value systems operate in parallel in the striatum for planning and extensively trained choices, with medial prefrontal cortex integrating their outputs.
Figures




References
-
- Sutton RS, Barto AG. Reinforcement Learning: An Introduction. MIT Press; Cambridge, MA: 1998.
-
- Samuels AL. Some studies in machine learning using the game of checkers. IBM Journal of Research and Development. 1957;3:210–229.
-
- O’Doherty JP, Dayan P, Friston K, Critchley H, Dolan RJ. Temporal difference models and reward-related learning in the human brain. Neuron. 2003;38:329–337. - PubMed
-
- Seymour B, et al. Temporal difference models describe higher-order learning in humans. Nature. 2004;429:664–667. - PubMed
-
- Shallice T. Specific impairments of planning. Philos Trans R Soc Lond B Biol Sci. 1982;298:199–209. - PubMed