Residential College | false |
Status | 已發表Published |
Breaking the Memory Wall for Heterogeneous Federated Learning via Model Splitting | |
Tian, Chunlin; Li, Li; Tam, Kahou; Wu, Yebo; Xu, Cheng Zhong | |
2024-12 | |
Source Publication | IEEE Transactions on Parallel and Distributed Systems |
ISSN | 1045-9219 |
Volume | 35Issue:12Pages:2513-2526 |
Abstract | Federated Learning (FL) enables multiple devices to collaboratively train a shared model while preserving data privacy. Ever-increasing model complexity coupled with limited memory resources on the participating devices severely bottlenecks the deployment of FL in real-world scenarios. Thus, a framework that can effectively break the memory wall while jointly taking into account the hardware and statistical heterogeneity in FL is urgently required. In this article, we propose SmartSplit a framework that effectively reduces the memory footprint on the device side while guaranteeing the training progress and model accuracy for heterogeneous FL through model splitting. Towards this end, SmartSplit employs a hierarchical structure to adaptively guide the overall training process. In each training round, the central manager, hosted on the server, dynamically selects the participating devices and sets the cutting layer by jointly considering the memory budget, training capacity, and data distribution of each device. The MEC manager, deployed within the edge server, proceeds to split the local model and perform training of the server-side portion. Meanwhile, it fine-tunes the splitting points based on the time-evolving statistical importance. The on-device manager, embedded inside each mobile device, continuously monitors the local training status while employing cost-aware checkpointing to match the runtime dynamic memory budget. Extensive experiments on representative datasets are conducted on both commercial off-the-shelf mobile device testbeds. The experimental results show that SmartSplit excels in FL training on highly memory-constrained mobile SoCs, offering up to a 94% peak latency reduction and 100-fold memory savings. It enhances accuracy performance by 1.49%-57.18% and adaptively adjusts to dynamic memory budgets through cost-aware recomputation. |
Keyword | Cross-device Federated Learning (Fl) Memory-wall Heterogeneity-aware |
DOI | 10.1109/TPDS.2024.3480115 |
URL | View the original |
Indexed By | SCIE |
Language | 英語English |
WOS Research Area | Computer Science ; Engineering |
WOS Subject | Computer Science, Theory & Methods ; Engineering, Electrical & Electronic |
WOS ID | WOS:001346106400001 |
Publisher | IEEE COMPUTER SOC, 10662 LOS VAQUEROS CIRCLE, PO BOX 3014, LOS ALAMITOS, CA 90720-1314 |
Scopus ID | 2-s2.0-85207726897 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | Faculty of Science and Technology THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU) DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE |
Corresponding Author | Li, Li |
Affiliation | State Key Laboratory of IoTSC, University of Macau, Taipa, Macau SAR, China |
First Author Affilication | University of Macau |
Corresponding Author Affilication | University of Macau |
Recommended Citation GB/T 7714 | Tian, Chunlin,Li, Li,Tam, Kahou,et al. Breaking the Memory Wall for Heterogeneous Federated Learning via Model Splitting[J]. IEEE Transactions on Parallel and Distributed Systems, 2024, 35(12), 2513-2526. |
APA | Tian, Chunlin., Li, Li., Tam, Kahou., Wu, Yebo., & Xu, Cheng Zhong (2024). Breaking the Memory Wall for Heterogeneous Federated Learning via Model Splitting. IEEE Transactions on Parallel and Distributed Systems, 35(12), 2513-2526. |
MLA | Tian, Chunlin,et al."Breaking the Memory Wall for Heterogeneous Federated Learning via Model Splitting".IEEE Transactions on Parallel and Distributed Systems 35.12(2024):2513-2526. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment