UM  > Faculty of Science and Technology
Residential Collegefalse
Status即將出版Forthcoming
Dealing with partial labels by knowledge distillation
Wang, Guangtai1; Huang, Jintao2; Lai, Yiqiang3; Vong, Chi Man1
2025-02-01
Source PublicationPattern Recognition
ISSN0031-3203
Volume158Pages:110965
Abstract

Partial label learning (PLL) is a weakly supervised methodology dealing with tasks that have annotation problems by replacing the single label with a collection of candidate labels. Compared to single labels, utilizing partial labels faces challenges: (1) The limited supervision and sensitivity to the false positive candidates; (2) Situations where the ground truth is not in the candidate label sets (noisy PLs). However, in the case that there exists a subset of samples that can be easily labeled (referred to as clean samples), the existing PLL paradigm needlessly assigns these instances with partial labels randomly. To better utilize the clean samples, and alleviate the obstacles of adopting partial labels, we proposed a specific Partial Label Knowledge Distillation (PLKD) framework to distill the knowledge from the samples with low annotating cost, further guiding partial label samples with limited supervision in these scenarios. The teacher model of PLKD was pre-trained on the clean samples with a single label, which can reduce the effect of noisy PLs when training on the remaining PLL samples. Additionally, recognizing that the existing candidate labels are sampled under the uniform distribution, which may not reflect real-life scenarios, we also proposed a label-specific candidate generation method. Correspondingly, a new loss function based on our generation method is presented to evaluate the distinction between partial labels and predictions. Furthermore, we also present a partial-label guided version, denoted as PLKD-pl, to alleviate the teacher's risk of over-confidence when the distribution between the clean set and partial label set varies widely. Extensive experimental evaluations have been conducted to demonstrate the superiority of PLKD over six state-of-the-art counterparts.

KeywordKnowledge Distillation Over-confidence Partial Label Learning
DOI10.1016/j.patcog.2024.110965
URLView the original
Indexed BySCIE
Language英語English
WOS Research AreaComputer Science ; Engineering
WOS SubjectComputer Science, Artificial Intelligence ; Engineering, Electrical & Electronic
WOS IDWOS:001309891000001
PublisherELSEVIER SCI LTD, 125 London Wall, London EC2Y 5AS, ENGLAND
Scopus ID2-s2.0-85203089011
Fulltext Access
Citation statistics
Document TypeJournal article
CollectionFaculty of Science and Technology
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorLai, Yiqiang
Affiliation1.Department of Computer and Information Science, University of Macau, China
2.Department of Computer Science, Hong Kong Baptist University, Hong Kong
3.South China Business College, Guangdong University of Foreign Studies, China
First Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Wang, Guangtai,Huang, Jintao,Lai, Yiqiang,et al. Dealing with partial labels by knowledge distillation[J]. Pattern Recognition, 2025, 158, 110965.
APA Wang, Guangtai., Huang, Jintao., Lai, Yiqiang., & Vong, Chi Man (2025). Dealing with partial labels by knowledge distillation. Pattern Recognition, 158, 110965.
MLA Wang, Guangtai,et al."Dealing with partial labels by knowledge distillation".Pattern Recognition 158(2025):110965.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Wang, Guangtai]'s Articles
[Huang, Jintao]'s Articles
[Lai, Yiqiang]'s Articles
Baidu academic
Similar articles in Baidu academic
[Wang, Guangtai]'s Articles
[Huang, Jintao]'s Articles
[Lai, Yiqiang]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Wang, Guangtai]'s Articles
[Huang, Jintao]'s Articles
[Lai, Yiqiang]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.