Residential College | false |
Status | 已發表Published |
Towards cost-effective service migration in mobile edge: A Q-learning approach | |
Wang,Yang1; Cao,Shan1; Ren,Hongshuai1; Li,Jianjun2; Ye,Kejiang1; Xu,Chengzhong3; Chen,Xi4 | |
2020-08-26 | |
Source Publication | Journal of Parallel and Distributed Computing |
ISSN | 0743-7315 |
Volume | 146Pages:175-188 |
Abstract | Service migration in mobile edge computing is a promising approach to improving the quality of service (QoS) for mobile users and reducing the network operational cost for service providers as well. However, these benefits are not free, coming at costs of bulk-data transfer, and likely service disruption, which could consequently increase the overall service costs. To gain the benefits of service migration while minimizing its cost across the edge nodes, in this paper, we leverage reinforcement learning (RL) method to design a cost-effective framework, called Mig-RL, for the service migration with a reduction of total service costs as a goal in a mobile edge environment. The Mig-RL leverages the infrastructure of edge network and deploys a migration agent through Q-learning to learn the optimal policy with respect to the service migration status. We distinguish the Mig-RL from other existing works in several major aspects. First, we fully exploit the nature of this problem in a modest migration space, which allows us to constrain the number of service replicas whereby a defined state–action space could be effectively handled, as opposed to those methods that need to always approximate a huge state–action space for policy optimality. Second, we advocate a migration policy-base as a cache to save the learning process by retrieving the most effective policy whenever a similar migration pattern is encountered as time goes on. Finally, by exploiting the idea of software defined network, we also investigate the efficient implementation of Mig-RL in mobile edge network. Experimental results based on some real and synthesized access sequences show that Mig-RL, compared with the selected existing algorithms, can substantially minimize the service costs, and in the meantime, efficiently improve the QoS by adapting to the changes of mobile access patterns. |
Keyword | Mobile Edge Computing Dynamic Service Migration Reinforcement Learning Q-learning Software-defined Networking |
DOI | 10.1016/j.jpdc.2020.08.008 |
URL | View the original |
Indexed By | SCIE |
Language | 英語English |
WOS Research Area | Computer Science |
WOS Subject | Computer Science, Theory & Methods |
WOS ID | WOS:000576604700007 |
Scopus ID | 2-s2.0-85090113588 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU) Faculty of Science and Technology |
Corresponding Author | Ye,Kejiang |
Affiliation | 1.Shenzhen Institutes of Advanced Technology,Chinese Academy of Sciences,518055,China 2.School of Computer Science and Technology,Hangzhou Dianzi University,Hangzhou,China 3.Faculty of Science and Technology,State Key Laboratory of IoT for Smart City,University of Macau,Macau SAR,China 4.CAS Research Center for Ecology and Environment of Central Asia,Urumqi,China |
Recommended Citation GB/T 7714 | Wang,Yang,Cao,Shan,Ren,Hongshuai,et al. Towards cost-effective service migration in mobile edge: A Q-learning approach[J]. Journal of Parallel and Distributed Computing, 2020, 146, 175-188. |
APA | Wang,Yang., Cao,Shan., Ren,Hongshuai., Li,Jianjun., Ye,Kejiang., Xu,Chengzhong., & Chen,Xi (2020). Towards cost-effective service migration in mobile edge: A Q-learning approach. Journal of Parallel and Distributed Computing, 146, 175-188. |
MLA | Wang,Yang,et al."Towards cost-effective service migration in mobile edge: A Q-learning approach".Journal of Parallel and Distributed Computing 146(2020):175-188. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment