Residential Collegefalse
Status即將出版Forthcoming
A fully value distributional deep reinforcement learning framework for multi-agent cooperation
Fu, Mingsheng1; Huang, Liwei1; Li, Fan2; Qu, Hong1; Xu, Chengzhong3
2025-04-01
Source PublicationNeural Networks
ISSN0893-6080
Volume184
Abstract

Distributional Reinforcement Learning (RL) extends beyond estimating the expected value of future returns by modeling its entire distribution, offering greater expressiveness and capturing deeper insights of the value function. To leverage this advantage, distributional multi-agent systems based on value-decomposition techniques were proposed recently. Ideally, a distributional multi-agent system should be fully distributional, which means both the individual and global value functions should be constructed in distributional forms. However, recent studies show that directly applying traditional value-decomposition techniques to this fully distributional form cannot guarantee the satisfaction of the necessary individual-global-max (IGM) principle. To address this problem, we propose a novel fully value distributional multi-agent framework based on value-decomposition and prove that the IGM principle can be guaranteed under our framework. Based on this framework, a practical deep reinforcement learning model called Fully Distributional Multi-Agent Cooperation (FDMAC) is proposed, and the effectiveness of FDMAC is verified under different scenarios of the StarCraft Multi-Agent Challenge micromanagement environment. Further experimental results show that our FDMAC model can outperform the best baseline by 10.47% on average in terms of the median test win rate.

KeywordDeep Reinforcement Learning Multi-agent Cooperation Distributional Reinforcement Learning Neural Networks
DOI10.1016/j.neunet.2024.107035
URLView the original
Indexed BySCIE
Language英語English
WOS Research AreaComputer Science ; Neurosciences & Neurology
WOS SubjectComputer Science, Artificial Intelligence ; Neurosciences
WOS IDWOS:001391001600001
PublisherPERGAMON-ELSEVIER SCIENCE LTD, THE BOULEVARD, LANGFORD LANE, KIDLINGTON, OXFORD OX5 1GB, ENGLAND
Scopus ID2-s2.0-85212192449
Fulltext Access
Citation statistics
Document TypeJournal article
CollectionTHE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU)
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorHuang, Liwei
Affiliation1.School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, Sichuan, 611731, China
2.Section of Epidemiology and Population Health, Department of Obstetrics and Gynecology, West China Second University Hospital, Sichuan University, Chengdu, Sichuan, 610041, China
3.State Key Laboratory of IoTSC, University of Macau, Taipa, 999078, Macao
Recommended Citation
GB/T 7714
Fu, Mingsheng,Huang, Liwei,Li, Fan,et al. A fully value distributional deep reinforcement learning framework for multi-agent cooperation[J]. Neural Networks, 2025, 184.
APA Fu, Mingsheng., Huang, Liwei., Li, Fan., Qu, Hong., & Xu, Chengzhong (2025). A fully value distributional deep reinforcement learning framework for multi-agent cooperation. Neural Networks, 184.
MLA Fu, Mingsheng,et al."A fully value distributional deep reinforcement learning framework for multi-agent cooperation".Neural Networks 184(2025).
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Fu, Mingsheng]'s Articles
[Huang, Liwei]'s Articles
[Li, Fan]'s Articles
Baidu academic
Similar articles in Baidu academic
[Fu, Mingsheng]'s Articles
[Huang, Liwei]'s Articles
[Li, Fan]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Fu, Mingsheng]'s Articles
[Huang, Liwei]'s Articles
[Li, Fan]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.