UM

Browse/Search Results:  1-10 of 13 Help

Selected(0)Clear Items/Page:    Sort:
Redundancy-free and load-balanced TGNN training with hierarchical pipeline parallelism Journal article
Xia, Yaqi, Zhang, Zheng, Yang, Donglin, Hu, Chuang, Zhou, Xiaobo, Chen, Hongyang, Sang, Qianlong, Cheng, Dazhao. Redundancy-free and load-balanced TGNN training with hierarchical pipeline parallelism[J]. IEEE Transactions on Parallel and Distributed Systems, 2024, 35(11), 1904-1919.
Authors:  Xia, Yaqi;  Zhang, Zheng;  Yang, Donglin;  Hu, Chuang;  Zhou, Xiaobo; et al.
Favorite | TC[WOS]:0 TC[Scopus]:0  IF:5.6/4.5 | Submit date:2024/08/05
Communication Balance  Distributed Training  Dynamic Gnn  Pipeline Parallelism  Redundancy-free  
DeepTM: Efficient Tensor Management in Heterogeneous Memory for DNN Training Journal article
Zhou, Haoran, Rang, Wei, Chen, Hongyang, Zhou, Xiaobo, Cheng, Dazhao. DeepTM: Efficient Tensor Management in Heterogeneous Memory for DNN Training[J]. IEEE Transactions on Parallel and Distributed Systems, 2024, 35(11), 1920-1935.
Authors:  Zhou, Haoran;  Rang, Wei;  Chen, Hongyang;  Zhou, Xiaobo;  Cheng, Dazhao
Favorite | TC[WOS]:0 TC[Scopus]:0  IF:5.6/4.5 | Submit date:2024/08/05
Deep Neural Network Training  Heterogeneous Memory  Memory Management  Performance Optimization  
Federated Spectrum Management Through Hedonic Coalition Formation Conference paper
He, Zhili, Tu, Tianyu, Wang, Kanye Ye, Luo, Bing, Cheng, Dazhao, Hu, Chuang. Federated Spectrum Management Through Hedonic Coalition Formation[C]:ASSOC COMPUTING MACHINERY1601 Broadway, 10th Floor, NEW YORK, NY, UNITED STATES, 2024, 217-218.
Authors:  He, Zhili;  Tu, Tianyu;  Wang, Kanye Ye;  Luo, Bing;  Cheng, Dazhao; et al.
Favorite | TC[WOS]:0 TC[Scopus]:0 | Submit date:2024/08/05
Raptor-T: A Fused and Memory-Efficient Sparse Transformer for Long and Variable-Length Sequences Journal article
Wang, Hulin, Yang, Donglin, Xia, Yaqi, Zhang, Zheng, Wang, Qigang, Fan, Jianping, Zhou, Xiaobo, Cheng, Dazhao. Raptor-T: A Fused and Memory-Efficient Sparse Transformer for Long and Variable-Length Sequences[J]. IEEE TRANSACTIONS ON COMPUTERS, 2024, 73(7), 1852-1865.
Authors:  Wang, Hulin;  Yang, Donglin;  Xia, Yaqi;  Zhang, Zheng;  Wang, Qigang; et al.
Favorite | TC[WOS]:1 TC[Scopus]:1  IF:3.6/3.2 | Submit date:2024/05/16
Sparse Transformer  Inference Acceleration  Gpu  Deep Learning  Memory Optimization  Resource Management  
Expeditious High-Concurrency MicroVM SnapStart in Persistent Memory with an Augmented Hypervisor Conference paper
XINGGUO PANG, YANZE ZHANG, LIU LIU, DAZHAO CHENG, CHENG-ZHONG XU, XIAOBO ZHOU. Expeditious High-Concurrency MicroVM SnapStart in Persistent Memory with an Augmented Hypervisor[C]:USENIX Association, 2024, 985-998.
Authors:  XINGGUO PANG;  YANZE ZHANG;  LIU LIU;  DAZHAO CHENG;  CHENG-ZHONG XU; et al.
Adobe PDF | Favorite | TC[WOS]:0 TC[Scopus]:0 | Submit date:2024/08/10
A unified hybrid memory system for scalable deep learning and big data applications Journal article
Rang, Wei, Liang, Huanghuang, Wang, Ye, Zhou, Xiaobo, Cheng, Dazhao. A unified hybrid memory system for scalable deep learning and big data applications[J]. Journal of Parallel and Distributed Computing, 2024, 186, 104820.
Authors:  Rang, Wei;  Liang, Huanghuang;  Wang, Ye;  Zhou, Xiaobo;  Cheng, Dazhao
Favorite | TC[WOS]:1 TC[Scopus]:1  IF:3.4/3.4 | Submit date:2024/05/02
Data Placement And Migration  Dnn Applications  Hybrid Memory System  Nvm  Unified Memory Management  
MPMoE: Memory Efficient MoE for Pre-Trained Models With Adaptive Pipeline Parallelism Journal article
Zhang, Zheng, Xia, Yaqi, Wang, Hulin, Yang, Donglin, Hu, Chuang, Zhou, Xiaobo, Cheng, Dazhao. MPMoE: Memory Efficient MoE for Pre-Trained Models With Adaptive Pipeline Parallelism[J]. IEEE Transactions on Parallel and Distributed Systems, 2024, 35(6), 843-856.
Authors:  Zhang, Zheng;  Xia, Yaqi;  Wang, Hulin;  Yang, Donglin;  Hu, Chuang; et al.
Favorite | TC[WOS]:0 TC[Scopus]:1  IF:5.6/4.5 | Submit date:2024/05/16
Distributed Training  Memory Redundancy  Mixture Of Experts  Performance Model  Pipeline Parallelism  
Locality-aware and Fault-tolerant Batching for Machine Learning on Distributed Datasets Journal article
Liu, Liu, Ding, Zhijun, Cheng, Dazhao, Zhou, Xiaobo. Locality-aware and Fault-tolerant Batching for Machine Learning on Distributed Datasets[J]. IEEE Transactions on Cloud Computing, 2024, 12(2), 370-387.
Authors:  Liu, Liu;  Ding, Zhijun;  Cheng, Dazhao;  Zhou, Xiaobo
Favorite | TC[WOS]:0 TC[Scopus]:0  IF:5.3/4.6 | Submit date:2024/05/16
Adaptation Models  Byzantine Gradient  Computational Modeling  Data Models  Distributed Databases  Distributed Dataset  Graphics Processing Units  Load Management  Machine Learning Training  Straggler  Training  
Incendio: Priority-based Scheduling for Alleviating Cold Start in Serverless Computing Journal article
Cai, Xinquan, Sang, Qianlong, Hu, Chuang, Gong, Yili, Suo, Kun, Zhou, Xiaobo, Cheng, Dazhao. Incendio: Priority-based Scheduling for Alleviating Cold Start in Serverless Computing[J]. IEEE Transactions on Computers, 2024, 73(7), 1780-1794.
Authors:  Cai, Xinquan;  Sang, Qianlong;  Hu, Chuang;  Gong, Yili;  Suo, Kun; et al.
Favorite | TC[WOS]:1 TC[Scopus]:1  IF:3.6/3.2 | Submit date:2024/05/16
Serverless Computing  Cold Start  Priority  Prediction  Scheduling  In-memory Computing  Distributed Systems  
Redundancy-Free High-Performance Dynamic GNN Training with Hierarchical Pipeline Parallelism Conference paper
Xia, Yaqi, Zhang, Zheng, Wang, Hulin, Yang, Donglin, Zhou, Xiaobo, Cheng, Dazhao. Redundancy-Free High-Performance Dynamic GNN Training with Hierarchical Pipeline Parallelism[C], 2023, 17-13.
Authors:  Xia, Yaqi;  Zhang, Zheng;  Wang, Hulin;  Yang, Donglin;  Zhou, Xiaobo; et al.
Favorite | TC[WOS]:4 TC[Scopus]:5 | Submit date:2023/08/08