Residential College | false |
Status | 已發表Published |
Collaborative Intelligent Reflecting Surface Networks With Multi-Agent Reinforcement Learning | |
Zhang, Jie1; Li, Jun1; Zhang, Yijin1; Wu, Qingqing2; Wu, Xiongwei3; Shu, Feng4; Jin, Shi5; Chen, Wen6 | |
2022-04-01 | |
Source Publication | IEEE Journal on Selected Topics in Signal Processing |
ISSN | 1932-4553 |
Volume | 16Issue:3Pages:532-545 |
Abstract | Intelligent reflecting surface (IRS) is envisioned to be widely applied in future wireless networks. In this paper, we investigate a multi-user communication system assisted by cooperative IRS devices with the capability of energy harvesting. Aiming to maximize the long-term average achievable system rate, an optimization problem is formulated by jointly designing the transmit beamforming at the base station (BS) and discrete phase shift beamforming at the IRSs, with the constraints on transmit power, user data rate requirement and IRS energy buffer size. Considering time-varying channels and stochastic arrivals of energy harvested by the IRSs, we first formulate the problem as a Markov decision process (MDP) and then develop a novel multi-agent Q-mix (MAQ) framework with two layers to decouple the optimization parameters. The higher layer is for optimizing phase shift resolutions, and the lower one is for phase shift beamforming and power allocation. Since the phase shift optimization is an integer programming problem with a large-scale action space, we improve MAQ by incorporating the Wolpertinger method, namely, MAQ-WP algorithm to achieve a sub-optimality with reduced dimensions of action space. In addition, as MAQ-WP is still of high complexity to achieve good performance, we propose a policy gradient-based MAQ algorithm, namely, MAQ-PG, by mapping the discrete phase shift actions into a continuous space at the cost of a slight performance loss. Simulation results demonstrate that the proposed MAQ-WP and MAQ-PG algorithms can converge faster and achieve data rate improvements of 10.7% and 8.8% over the conventional multi-agent DDPG, respectively. |
Keyword | Beamforming Energy Harvesting Intelligent Reflecting Surface Multi-agent Reinforcement Learning |
DOI | 10.1109/JSTSP.2022.3162109 |
URL | View the original |
Indexed By | SCIE |
Language | 英語English |
WOS Research Area | Engineering |
WOS Subject | Engineering, Electrical & Electronic |
WOS ID | WOS:000797421100021 |
Scopus ID | 2-s2.0-85130700435 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU) |
Corresponding Author | Li, Jun |
Affiliation | 1.Nanjing University of Science and Technology, School of Electronic and Optical Engineering, Nanjing, 210094, China 2.University of Macau, State Key Laboratory of IoT for Smart City, 999078, Macao 3.Chinese University of Hong Kong, Department of Electronic Engineering, Hong Kong, Hong Kong 4.Hainan University, School of Information and Communication Engineering, Haikou, 570228, China 5.Southeast University, National Mobile Communications Research Laboratory, Nanjing, 210096, China 6.Shanghai Jiao Tong University, Department of Electronic Engineering, Shanghai, 200240, China |
Recommended Citation GB/T 7714 | Zhang, Jie,Li, Jun,Zhang, Yijin,et al. Collaborative Intelligent Reflecting Surface Networks With Multi-Agent Reinforcement Learning[J]. IEEE Journal on Selected Topics in Signal Processing, 2022, 16(3), 532-545. |
APA | Zhang, Jie., Li, Jun., Zhang, Yijin., Wu, Qingqing., Wu, Xiongwei., Shu, Feng., Jin, Shi., & Chen, Wen (2022). Collaborative Intelligent Reflecting Surface Networks With Multi-Agent Reinforcement Learning. IEEE Journal on Selected Topics in Signal Processing, 16(3), 532-545. |
MLA | Zhang, Jie,et al."Collaborative Intelligent Reflecting Surface Networks With Multi-Agent Reinforcement Learning".IEEE Journal on Selected Topics in Signal Processing 16.3(2022):532-545. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment