4.6 Article

FMRQ-A Multiagent Reinforcement Learning Algorithm for Fully Cooperative Tasks

期刊

IEEE TRANSACTIONS ON CYBERNETICS
卷 47, 期 6, 页码 1367-1379

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TCYB.2016.2544866

关键词

Multiagent reinforcement learning (MARL); Nash equilibrium; Q-learning; repeated game

资金

  1. National Natural Science Foundation of China [61273136, 61573353, 61533017, 61573205]
  2. Foundation of Shandong Province [ZR2015FM015, ZR2015FM017]

向作者/读者索取更多资源

In this paper, we propose a multiagent reinforcement learning algorithm dealing with fully cooperative tasks. The algorithm is called frequency of the maximum reward Q-learning (FMRQ). FMRQ aims to achieve one of the optimal Nash equilibria so as to optimize the performance index in multiagent systems. The frequency of obtaining the highest global immediate reward instead of immediate reward is used as the reinforcement signal. With FMRQ each agent does not need the observation of the other agents' actions and only shares its state and reward at each step. We validate FMRQ through case studies of repeated games: four cases of two-player two-action and one case of three-player two-action. It is demonstrated that FMRQ can converge to one of the optimal Nash equilibria in these cases. Moreover, comparison experiments on tasks with multiple states and finite steps are conducted. One is box-pushing and the other one is distributed sensor network problem. Experimental results show that the proposed algorithm outperforms others with higher performance.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据