UM

Browse/Search Results:  1-10 of 13 Help

Selected(0)Clear Items/Page:    Sort:
DeepTM: Efficient Tensor Management in Heterogeneous Memory for DNN Training Journal article
Zhou, Haoran, Rang, Wei, Chen, Hongyang, Zhou, Xiaobo, Cheng, Dazhao. DeepTM: Efficient Tensor Management in Heterogeneous Memory for DNN Training[J]. IEEE Transactions on Parallel and Distributed Systems, 2024, 35(11), 1920-1935.
Authors:  Zhou, Haoran;  Rang, Wei;  Chen, Hongyang;  Zhou, Xiaobo;  Cheng, Dazhao
Favorite | TC[WOS]:0 TC[Scopus]:0  IF:5.6/4.5 | Submit date:2024/08/05
Deep Neural Network Training  Heterogeneous Memory  Memory Management  Performance Optimization  
Redundancy-free and load-balanced TGNN training with hierarchical pipeline parallelism Journal article
Xia, Yaqi, Zhang, Zheng, Yang, Donglin, Hu, Chuang, Zhou, Xiaobo, Chen, Hongyang, Sang, Qianlong, Cheng, Dazhao. Redundancy-free and load-balanced TGNN training with hierarchical pipeline parallelism[J]. IEEE Transactions on Parallel and Distributed Systems, 2024, 35(11), 1904-1919.
Authors:  Xia, Yaqi;  Zhang, Zheng;  Yang, Donglin;  Hu, Chuang;  Zhou, Xiaobo; et al.
Favorite | TC[WOS]:0 TC[Scopus]:0  IF:5.6/4.5 | Submit date:2024/08/05
Communication Balance  Distributed Training  Dynamic Gnn  Pipeline Parallelism  Redundancy-free  
Federated Spectrum Management Through Hedonic Coalition Formation Conference paper
He, Zhili, Tu, Tianyu, Wang, Kanye Ye, Luo, Bing, Cheng, Dazhao, Hu, Chuang. Federated Spectrum Management Through Hedonic Coalition Formation[C]:ASSOC COMPUTING MACHINERY1601 Broadway, 10th Floor, NEW YORK, NY, UNITED STATES, 2024, 217-218.
Authors:  He, Zhili;  Tu, Tianyu;  Wang, Kanye Ye;  Luo, Bing;  Cheng, Dazhao; et al.
Favorite | TC[WOS]:0 TC[Scopus]:0 | Submit date:2024/08/05
Expeditious High-Concurrency MicroVM SnapStart in Persistent Memory with an Augmented Hypervisor Conference paper
XINGGUO PANG, YANZE ZHANG, LIU LIU, DAZHAO CHENG, CHENG-ZHONG XU, XIAOBO ZHOU. Expeditious High-Concurrency MicroVM SnapStart in Persistent Memory with an Augmented Hypervisor[C]:USENIX Association, 2024, 985-998.
Authors:  XINGGUO PANG;  YANZE ZHANG;  LIU LIU;  DAZHAO CHENG;  CHENG-ZHONG XU; et al.
Adobe PDF | Favorite | TC[WOS]:0 TC[Scopus]:0 | Submit date:2024/08/10
Raptor-T: A Fused and Memory-Efficient Sparse Transformer for Long and Variable-Length Sequences Journal article
Wang, Hulin, Yang, Donglin, Xia, Yaqi, Zhang, Zheng, Wang, Qigang, Fan, Jianping, Zhou, Xiaobo, Cheng, Dazhao. Raptor-T: A Fused and Memory-Efficient Sparse Transformer for Long and Variable-Length Sequences[J]. IEEE TRANSACTIONS ON COMPUTERS, 2024, 73(7), 1852-1865.
Authors:  Wang, Hulin;  Yang, Donglin;  Xia, Yaqi;  Zhang, Zheng;  Wang, Qigang; et al.
Favorite | TC[WOS]:1 TC[Scopus]:1  IF:3.6/3.2 | Submit date:2024/05/16
Sparse Transformer  Inference Acceleration  Gpu  Deep Learning  Memory Optimization  Resource Management  
A unified hybrid memory system for scalable deep learning and big data applications Journal article
Rang, Wei, Liang, Huanghuang, Wang, Ye, Zhou, Xiaobo, Cheng, Dazhao. A unified hybrid memory system for scalable deep learning and big data applications[J]. Journal of Parallel and Distributed Computing, 2024, 186, 104820.
Authors:  Rang, Wei;  Liang, Huanghuang;  Wang, Ye;  Zhou, Xiaobo;  Cheng, Dazhao
Favorite | TC[WOS]:1 TC[Scopus]:1  IF:3.4/3.4 | Submit date:2024/05/02
Data Placement And Migration  Dnn Applications  Hybrid Memory System  Nvm  Unified Memory Management  
MPMoE: Memory Efficient MoE for Pre-Trained Models With Adaptive Pipeline Parallelism Journal article
Zhang, Zheng, Xia, Yaqi, Wang, Hulin, Yang, Donglin, Hu, Chuang, Zhou, Xiaobo, Cheng, Dazhao. MPMoE: Memory Efficient MoE for Pre-Trained Models With Adaptive Pipeline Parallelism[J]. IEEE Transactions on Parallel and Distributed Systems, 2024, 35(6), 843-856.
Authors:  Zhang, Zheng;  Xia, Yaqi;  Wang, Hulin;  Yang, Donglin;  Hu, Chuang; et al.
Favorite | TC[WOS]:0 TC[Scopus]:1  IF:5.6/4.5 | Submit date:2024/05/16
Distributed Training  Memory Redundancy  Mixture Of Experts  Performance Model  Pipeline Parallelism  
Locality-aware and Fault-tolerant Batching for Machine Learning on Distributed Datasets Journal article
Liu, Liu, Ding, Zhijun, Cheng, Dazhao, Zhou, Xiaobo. Locality-aware and Fault-tolerant Batching for Machine Learning on Distributed Datasets[J]. IEEE Transactions on Cloud Computing, 2024, 12(2), 370-387.
Authors:  Liu, Liu;  Ding, Zhijun;  Cheng, Dazhao;  Zhou, Xiaobo
Favorite | TC[WOS]:0 TC[Scopus]:0  IF:5.3/4.6 | Submit date:2024/05/16
Adaptation Models  Byzantine Gradient  Computational Modeling  Data Models  Distributed Databases  Distributed Dataset  Graphics Processing Units  Load Management  Machine Learning Training  Straggler  Training  
Incendio: Priority-based Scheduling for Alleviating Cold Start in Serverless Computing Journal article
Cai, Xinquan, Sang, Qianlong, Hu, Chuang, Gong, Yili, Suo, Kun, Zhou, Xiaobo, Cheng, Dazhao. Incendio: Priority-based Scheduling for Alleviating Cold Start in Serverless Computing[J]. IEEE Transactions on Computers, 2024, 73(7), 1780-1794.
Authors:  Cai, Xinquan;  Sang, Qianlong;  Hu, Chuang;  Gong, Yili;  Suo, Kun; et al.
Favorite | TC[WOS]:1 TC[Scopus]:1  IF:3.6/3.2 | Submit date:2024/05/16
Serverless Computing  Cold Start  Priority  Prediction  Scheduling  In-memory Computing  Distributed Systems  
Redundancy-Free High-Performance Dynamic GNN Training with Hierarchical Pipeline Parallelism Conference paper
Xia, Yaqi, Zhang, Zheng, Wang, Hulin, Yang, Donglin, Zhou, Xiaobo, Cheng, Dazhao. Redundancy-Free High-Performance Dynamic GNN Training with Hierarchical Pipeline Parallelism[C], 2023, 17-13.
Authors:  Xia, Yaqi;  Zhang, Zheng;  Wang, Hulin;  Yang, Donglin;  Zhou, Xiaobo; et al.
Favorite | TC[WOS]:4 TC[Scopus]:5 | Submit date:2023/08/08