Residential College | false |
Status | 即將出版Forthcoming |
A fully value distributional deep reinforcement learning framework for multi-agent cooperation | |
Fu, Mingsheng1; Huang, Liwei1![]() ![]() | |
2025-04-01 | |
Source Publication | Neural Networks
![]() |
ISSN | 0893-6080 |
Volume | 184 |
Abstract | Distributional Reinforcement Learning (RL) extends beyond estimating the expected value of future returns by modeling its entire distribution, offering greater expressiveness and capturing deeper insights of the value function. To leverage this advantage, distributional multi-agent systems based on value-decomposition techniques were proposed recently. Ideally, a distributional multi-agent system should be fully distributional, which means both the individual and global value functions should be constructed in distributional forms. However, recent studies show that directly applying traditional value-decomposition techniques to this fully distributional form cannot guarantee the satisfaction of the necessary individual-global-max (IGM) principle. To address this problem, we propose a novel fully value distributional multi-agent framework based on value-decomposition and prove that the IGM principle can be guaranteed under our framework. Based on this framework, a practical deep reinforcement learning model called Fully Distributional Multi-Agent Cooperation (FDMAC) is proposed, and the effectiveness of FDMAC is verified under different scenarios of the StarCraft Multi-Agent Challenge micromanagement environment. Further experimental results show that our FDMAC model can outperform the best baseline by 10.47% on average in terms of the median test win rate. |
Keyword | Deep Reinforcement Learning Multi-agent Cooperation Distributional Reinforcement Learning Neural Networks |
DOI | 10.1016/j.neunet.2024.107035 |
URL | View the original |
Indexed By | SCIE |
Language | 英語English |
WOS Research Area | Computer Science ; Neurosciences & Neurology |
WOS Subject | Computer Science, Artificial Intelligence ; Neurosciences |
WOS ID | WOS:001391001600001 |
Publisher | PERGAMON-ELSEVIER SCIENCE LTD, THE BOULEVARD, LANGFORD LANE, KIDLINGTON, OXFORD OX5 1GB, ENGLAND |
Scopus ID | 2-s2.0-85212192449 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU) DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE |
Corresponding Author | Huang, Liwei |
Affiliation | 1.School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, Sichuan, 611731, China 2.Section of Epidemiology and Population Health, Department of Obstetrics and Gynecology, West China Second University Hospital, Sichuan University, Chengdu, Sichuan, 610041, China 3.State Key Laboratory of IoTSC, University of Macau, Taipa, 999078, Macao |
Recommended Citation GB/T 7714 | Fu, Mingsheng,Huang, Liwei,Li, Fan,et al. A fully value distributional deep reinforcement learning framework for multi-agent cooperation[J]. Neural Networks, 2025, 184. |
APA | Fu, Mingsheng., Huang, Liwei., Li, Fan., Qu, Hong., & Xu, Chengzhong (2025). A fully value distributional deep reinforcement learning framework for multi-agent cooperation. Neural Networks, 184. |
MLA | Fu, Mingsheng,et al."A fully value distributional deep reinforcement learning framework for multi-agent cooperation".Neural Networks 184(2025). |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment