4.3 Article

Learning to represent reward structure: A key to adapting to complex environments

期刊

NEUROSCIENCE RESEARCH
卷 74, 期 3-4, 页码 177-183

出版社

ELSEVIER IRELAND LTD
DOI: 10.1016/j.neures.2012.09.007

关键词

Reward; Dopamine; Reinforcement learning; Decision; Value; Salience; Structure

资金

  1. KAKENHI [21300129, 24120522]
  2. Grants-in-Aid for Scientific Research [24120523, 21300129, 24120522] Funding Source: KAKEN

向作者/读者索取更多资源

Predicting outcomes is a critical ability of humans and animals. The dopamine reward prediction error hypothesis, the driving force behind the recent progress in neural value-based decision making, states that dopamine activity encodes the signals for learning in order to predict a reward, that is, the difference between the actual and predicted reward, called the reward prediction error. However, this hypothesis and its underlying assumptions limit the prediction and its error as reactively triggered by momentary environmental events. Reviewing the assumptions and some of the latest findings, we suggest that the internal state representation is learned to reflect the environmental reward structure, and we propose a new hypothesis - the dopamine reward structural learning hypothesis - in which dopamine activity encodes multiplex signals for learning in order to represent reward structure in the internal state, leading to better reward prediction. (C) 2012 Elsevier Ireland Ltd and the Japan Neuroscience Society. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.3
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据