Pessiglione, M and Seymour, B and Flandin, G and Dolan, RJ and Frith, CD (2006) Dopamine-dependent prediction errors underpin reward-seeking behaviour in humans. Nature, 442. pp. 1042-1045.Full text not available from this repository.
Theories of instrumental learning are centred on understanding how success and failure are used to improve future decisions. These theories highlight a central role for reward prediction errors in updating the values associated with available actions. In animals, substantial evidence indicates that the neurotransmitter dopamine might have a key function in this type of learning, through its ability to modulate cortico-striatal synaptic efficacy. However, no direct evidence links dopamine, striatal activity and behavioural choice in humans. Here we show that, during instrumental learning, the magnitude of reward prediction error expressed in the striatum is modulated by the administration of drugs enhancing (3,4-dihydroxy-L-phenylalanine; L-DOPA) or reducing (haloperidol) dopaminergic function. Accordingly, subjects treated with L-DOPA have a greater propensity to choose the most rewarding action relative to subjects treated with haloperidol. Furthermore, incorporating the magnitude of the prediction errors into a standard action-value learning algorithm accurately reproduced subjects' behavioural choices under the different drug conditions. We conclude that dopamine-dependent modulation of striatal activity can account for how the human brain uses reward prediction errors to improve future decisions.
|Uncontrolled Keywords:||Adult Algorithms Behavior Computer Simulation Computers Dopamine Dopamine Agents Dopamine Antagonists Female Forecasting Haloperidol Humans Learning Levodopa Male Models, Neurological Neostriatum Punishment Reward|
|Divisions:||Div F > Computational and Biological Learning|
|Depositing User:||Unnamed user with email email@example.com|
|Date Deposited:||16 Jul 2015 13:17|
|Last Modified:||04 Sep 2015 00:31|