UM  > Faculty of Science and Technology
Residential Collegefalse
Status已發表Published
Federated learning on non-IID and globally long-tailed data via meta re-weighting networks
Lu, Yang1; Qian, Pinxin1; Yan, Shanshan1; Huang, Gang2; Tang, Yuan Yan3
2024-05-01
Source PublicationInternational Journal of Wavelets, Multiresolution and Information Processing
ISSN0219-6913
Volume22Issue:3Pages:2350063
Abstract

Under a federated learning environment, the training samples are generally collected and stored locally on each client's device, which makes the machine learning procedure not meet the requirement of independent and identical distribution (IID). Existing federated learning methods to deal with non-IID data generally assume that the data is globally balanced. However, real-world multi-class data tend to exhibit long-tail distribution, where the majority of samples are in a few head classes and a large number of tail classes only have a small amount of data. This paper, therefore, focuses on addressing the problem of handling non-IID and globally long-tailed data in a federated learning scenario. Accordingly, we propose a new federated learning method called Federated meta re-weighting networks (FedReN), which assigns weights during the local training process from the class-level and instance-level perspectives, respectively. To deal with data non-IIDness and global long-tail, both of the two re-weighting functions are globally trained by the meta-learning approach to acquire the knowledge of global long-tail distribution. Experiments on several long-tailed image classification benchmarks show that FedReN outperforms the state-of-the-art federated learning methods. The code is available at https://github.com/pxqian/FedReN.

KeywordFederated Learning Long-tail Learning Meta-learning Re-weighting
DOI10.1142/S0219691323500637
URLView the original
Indexed BySCIE
Language英語English
WOS Research AreaComputer Science ; Mathematics
WOS SubjectComputer Science, Software Engineering ; Mathematics, Interdisciplinary Applications
WOS IDWOS:001137341500001
PublisherWORLD SCIENTIFIC PUBL CO PTE LTD, 5 TOH TUCK LINK, SINGAPORE 596224, SINGAPORE
Scopus ID2-s2.0-85181969654
Fulltext Access
Citation statistics
Document TypeJournal article
CollectionFaculty of Science and Technology
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorTang, Yuan Yan
Affiliation1.School of Informatics, Xiamen University, Xiamen, China
2.College of Electrical Engineering, Zhejiang University, Hangzhou, China
3.Zhuhai UM Science and Technology Research Institute, University of Macau, Macau, Macao
Corresponding Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Lu, Yang,Qian, Pinxin,Yan, Shanshan,et al. Federated learning on non-IID and globally long-tailed data via meta re-weighting networks[J]. International Journal of Wavelets, Multiresolution and Information Processing, 2024, 22(3), 2350063.
APA Lu, Yang., Qian, Pinxin., Yan, Shanshan., Huang, Gang., & Tang, Yuan Yan (2024). Federated learning on non-IID and globally long-tailed data via meta re-weighting networks. International Journal of Wavelets, Multiresolution and Information Processing, 22(3), 2350063.
MLA Lu, Yang,et al."Federated learning on non-IID and globally long-tailed data via meta re-weighting networks".International Journal of Wavelets, Multiresolution and Information Processing 22.3(2024):2350063.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Lu, Yang]'s Articles
[Qian, Pinxin]'s Articles
[Yan, Shanshan]'s Articles
Baidu academic
Similar articles in Baidu academic
[Lu, Yang]'s Articles
[Qian, Pinxin]'s Articles
[Yan, Shanshan]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Lu, Yang]'s Articles
[Qian, Pinxin]'s Articles
[Yan, Shanshan]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.