UM

Browse/Search Results:  1-1 of 1 Help

Selected(0)Clear Items/Page:    Sort:
UELLM: A Unified and Efficient Approach for Large Language Model Inference Serving Conference paper
He, Yiyuan, Xu, Minxian, Wu, Jingfeng, Zheng, Wanyi, Ye, Kejiang, Xu, Chengzhong. UELLM: A Unified and Efficient Approach for Large Language Model Inference Serving[C]:Springer Science and Business Media Deutschland GmbH, 2025, 218-235.
Authors:  He, Yiyuan;  Xu, Minxian;  Wu, Jingfeng;  Zheng, Wanyi;  Ye, Kejiang; et al.
Favorite | TC[Scopus]:0 | Submit date:2025/01/22
Cloud Computing  Large Language Model Inference  Resource Management  Scheduling Algorithm