Residential College | false |
Status | 已發表Published |
FedHybrid: Breaking the Memory Wall of Federated Learning via Hybrid Tensor Management | |
Tam, Kahou1; Tian, Chunlin1; Li, Li1; Zhao, Haikai2; Xu, Cheng Zhong1 | |
2024-11-04 | |
Conference Name | SenSys '24: 22nd ACM Conference on Embedded Networked Sensor Systems |
Source Publication | SenSys 2024 - Proceedings of the 2024 ACM Conference on Embedded Networked Sensor Systems |
Pages | 394-408 |
Conference Date | November 4-7 2024 |
Conference Place | Hangzhou, China |
Country | CHINA |
Publication Place | New York, NY, USA |
Publisher | Association for Computing Machinery |
Abstract | Federated Learning (FL) emerges as a new learning paradigm that enables multiple devices to collaboratively train a shared model while preserving data privacy. However, one fundamental and prevailing challenge that hinders the deployment of FL on mobile devices is the memory limitation. This paper proposes FedHybrid, a novel framework that effectively reduces the memory footprint during the training process while guaranteeing the model accuracy and the overall training progress. Specifically, FedHybrid first selects the participating devices for each training round by jointly evaluating their memory budget, computing capability, and data diversity. After that, it judiciously analyzes the computational graph and generates an execution plan for each selected client in order to meet the corresponding memory budget while minimizing the training delay through employing a hybrid of recomputation and compression techniques according to the characteristic of each tensor. During the local training process, FedHybrid carries out the execution plan with a well-designed activation compression technique to effectively achieve memory reduction with minimum accuracy loss. We conduct extensive experiments to evaluate FedHybrid on both simulation and off-the-shelf mobile devices. The experiment results demonstrate that FedHybrid achieves up to a 39.1% increase in model accuracy and a 15.5X reduction in wall clock time under various memory budgets compared with the baselines. |
Keyword | Federated Learning Memory Optimization Mobile Computing |
DOI | 10.1145/3666025.3699346 |
URL | View the original |
Language | 英語English |
Scopus ID | 2-s2.0-85211782514 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | Faculty of Science and Technology THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU) DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE |
Corresponding Author | Li, Li |
Affiliation | 1.State Key Laboratory of IoTSC, University of Macau, Macau, Macao 2.Simon Fraser University, Vancouver, Canada |
First Author Affilication | University of Macau |
Corresponding Author Affilication | University of Macau |
Recommended Citation GB/T 7714 | Tam, Kahou,Tian, Chunlin,Li, Li,et al. FedHybrid: Breaking the Memory Wall of Federated Learning via Hybrid Tensor Management[C], New York, NY, USA:Association for Computing Machinery, 2024, 394-408. |
APA | Tam, Kahou., Tian, Chunlin., Li, Li., Zhao, Haikai., & Xu, Cheng Zhong (2024). FedHybrid: Breaking the Memory Wall of Federated Learning via Hybrid Tensor Management. SenSys 2024 - Proceedings of the 2024 ACM Conference on Embedded Networked Sensor Systems, 394-408. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment