×
验证码:
换一张
Forgotten Password?
Stay signed in
Login With UMPASS
English
|
繁體
Login With UMPASS
Log In
ALL
ORCID
TI
AU
PY
SU
KW
TY
JN
DA
IN
PB
FP
ST
SM
Study Hall
Image search
Paste the image URL
Home
Faculties & Institutes
Scholars
Publications
Subjects
Statistics
News
Search in the results
Faculties & Institutes
Faculty of Scien... [2]
THE STATE KEY LA... [1]
Authors
CHENGZHONG XU [1]
XIAOBO ZHOU [1]
Document Type
Journal article [2]
Date Issued
2024 [2]
Language
英語English [2]
Source Publication
IEEE TRANSACTION... [1]
IEEE Transaction... [1]
Indexed By
SCIE [2]
Funding Organization
Funding Project
×
Knowledge Map
UM
Start a Submission
Submissions
Unclaimed
Claimed
Attach Fulltext
Bookmarks
Browse/Search Results:
1-2 of 2
Help
Selected(
0
)
Clear
Items/Page:
5
10
15
20
25
30
35
40
45
50
55
60
65
70
75
80
85
90
95
100
Sort:
Select
Issue Date Ascending
Issue Date Descending
Journal Impact Factor Ascending
Journal Impact Factor Descending
WOS Cited Times Ascending
WOS Cited Times Descending
Submit date Ascending
Submit date Descending
Title Ascending
Title Descending
Author Ascending
Author Descending
InSS: An Intelligent Scheduling Orchestrator for Multi-GPU Inference with Spatio-Temporal Sharing
Journal article
Han, Ziyi, Zhou, Ruiting, Xu, Chengzhong, Zeng, Yifan, Zhang, Renli. InSS: An Intelligent Scheduling Orchestrator for Multi-GPU Inference with Spatio-Temporal Sharing[J]. IEEE Transactions on Parallel and Distributed Systems, 2024, 35(10), 1735-1748.
Authors:
Han, Ziyi
;
Zhou, Ruiting
;
Xu, Chengzhong
;
Zeng, Yifan
;
Zhang, Renli
Favorite
|
TC[WOS]:
0
TC[Scopus]:
2
IF:
5.6
/
4.5
|
Submit date:2024/08/05
Dnn Inference
Gpu Resource Management
Online Scheduling
Raptor-T: A Fused and Memory-Efficient Sparse Transformer for Long and Variable-Length Sequences
Journal article
Wang, Hulin, Yang, Donglin, Xia, Yaqi, Zhang, Zheng, Wang, Qigang, Fan, Jianping, Zhou, Xiaobo, Cheng, Dazhao. Raptor-T: A Fused and Memory-Efficient Sparse Transformer for Long and Variable-Length Sequences[J]. IEEE TRANSACTIONS ON COMPUTERS, 2024, 73(7), 1852-1865.
Authors:
Wang, Hulin
;
Yang, Donglin
;
Xia, Yaqi
;
Zhang, Zheng
;
Wang, Qigang
; et al.
Favorite
|
TC[WOS]:
1
TC[Scopus]:
1
IF:
3.6
/
3.2
|
Submit date:2024/05/16
Sparse Transformer
Inference Acceleration
Gpu
Deep Learning
Memory Optimization
Resource Management