The ubiquity of model-based reinforcement learning.

Curr Opin Neurobiol

Department of Psychology, Columbia University, New York, NY, United States.

Published: December 2012


Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

The reward prediction error (RPE) theory of dopamine (DA) function has enjoyed great success in the neuroscience of learning and decision-making. This theory is derived from model-free reinforcement learning (RL), in which choices are made simply on the basis of previously realized rewards. Recently, attention has turned to correlates of more flexible, albeit computationally complex, model-based methods in the brain. These methods are distinguished from model-free learning by their evaluation of candidate actions using expected future outcomes according to a world model. Puzzlingly, signatures from these computations seem to be pervasive in the very same regions previously thought to support model-free learning. Here, we review recent behavioral and neural evidence about these two systems, in attempt to reconcile their enigmatic cohabitation in the brain.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC3513648PMC
http://dx.doi.org/10.1016/j.conb.2012.08.003DOI Listing

Publication Analysis

Top Keywords

reinforcement learning
8
model-free learning
8
learning
5
ubiquity model-based
4
model-based reinforcement
4
learning reward
4
reward prediction
4
prediction error
4
error rpe
4
rpe theory
4

Similar Publications