UM  > Faculty of Science and Technology
Residential Collegefalse
Status已發表Published
MPMoE: Memory Efficient MoE for Pre-Trained Models With Adaptive Pipeline Parallelism
Zhang, Zheng1; Xia, Yaqi1; Wang, Hulin1; Yang, Donglin2; Hu, Chuang1; Zhou, Xiaobo3; Cheng, Dazhao1
2024-04
Source PublicationIEEE Transactions on Parallel and Distributed Systems
ISSN1045-9219
Volume35Issue:6Pages:843-856
Abstract

In recent years, the Mixture-of-Experts (MoE) technique has gained widespread popularity as a means to scale pre-trained models to exceptionally large sizes. Dynamic activation of experts allows for conditional computation, increasing the number of parameters of neural networks, which is critical for absorbing the vast amounts of knowledge available in many deep learning areas. However, despite the existing system and algorithm optimizations, there are significant challenges to be tackled when it comes to the inefficiencies of communication and memory consumption. In this paper, we present the design and implementation of MPMoE, a high-performance library that accelerates MoE training with adaptive and memory-efficient pipeline parallelism. Inspired by that the MoE training procedure can be divided into multiple independent sub-stages. We design a pipeline parallelism method for reducing communication latency by overlapping with computation operations. Further, we analyze the memory footprint breakdown of MoE training and identify that activations and temporary buffers are the primary contributors to the overall memory footprint. Toward memory efficiency, we propose memory reuse strategies to reduce memory requirements by eliminating memory redundancies. Finally, to optimize pipeline granularity and memory reuse strategies jointly, we propose a profile-based algorithm and a performance model to determine the configurations of MPMoE at runtime. We implement MPMoE upon PyTorch and evaluate it with common MoE models in two physical clusters, including 64 NVIDIA A100 GPU cards and 16 NVIDIA V100 GPU cards. Compared with the state-of-art approach, MPMoE achieves up to 2.3× speedup while reducing more than 30% memory footprint for training large models.

KeywordDistributed Training Memory Redundancy Mixture Of Experts Performance Model Pipeline Parallelism
DOI10.1109/TPDS.2024.3385639
URLView the original
Indexed BySCIE
Language英語English
WOS Research AreaComputer Science ; Engineering
WOS SubjectComputer Science, Theory & Methods ; Engineering, Electrical & Electronic
WOS IDWOS:001209556500002
PublisherIEEE COMPUTER SOC, 10662 LOS VAQUEROS CIRCLE, PO BOX 3014, LOS ALAMITOS, CA 90720-1314
Scopus ID2-s2.0-85190172633
Fulltext Access
Citation statistics
Document TypeJournal article
CollectionFaculty of Science and Technology
THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU)
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorCheng, Dazhao
Affiliation1.Wuhan University, School of Computer Science, Hubei, 430072, China
2.Nvidia Corp., Santa Clara, 95051, United States
3.University of Macau, Iotsc and Department of Computer and Information Sciences, Macao
Recommended Citation
GB/T 7714
Zhang, Zheng,Xia, Yaqi,Wang, Hulin,et al. MPMoE: Memory Efficient MoE for Pre-Trained Models With Adaptive Pipeline Parallelism[J]. IEEE Transactions on Parallel and Distributed Systems, 2024, 35(6), 843-856.
APA Zhang, Zheng., Xia, Yaqi., Wang, Hulin., Yang, Donglin., Hu, Chuang., Zhou, Xiaobo., & Cheng, Dazhao (2024). MPMoE: Memory Efficient MoE for Pre-Trained Models With Adaptive Pipeline Parallelism. IEEE Transactions on Parallel and Distributed Systems, 35(6), 843-856.
MLA Zhang, Zheng,et al."MPMoE: Memory Efficient MoE for Pre-Trained Models With Adaptive Pipeline Parallelism".IEEE Transactions on Parallel and Distributed Systems 35.6(2024):843-856.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Zhang, Zheng]'s Articles
[Xia, Yaqi]'s Articles
[Wang, Hulin]'s Articles
Baidu academic
Similar articles in Baidu academic
[Zhang, Zheng]'s Articles
[Xia, Yaqi]'s Articles
[Wang, Hulin]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Zhang, Zheng]'s Articles
[Xia, Yaqi]'s Articles
[Wang, Hulin]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.