Residential College | false |
Status | 已發表Published |
Generalized Few-Shot Node Classification With Graph Knowledge Distillation | |
Wang, Jialong; Zhou, Mengting; Zhang, Shilong; Gong, Zhiguo![]() ![]() | |
2024-04-18 | |
Source Publication | IEEE Transactions on Computational Social Systems
![]() |
ISSN | 2329-924X |
Abstract | Generalized few-shot node classification (GFS-NC) is a very important challenge for graph-based algorithms, as it requires to identify novel classes and base classes simultaneously. Although there are several methods that try to combine metalearning or metric learning with graph neural networks to solve few-shot problem, most of them assume that test samples only come from the novel classes, which is impractical in reality. Besides, they overlook the relationship among classes, which can provide additional information for the novel classes classification. In this article, we propose a graph-based knowledge distillation network (GraphKD) to extract the class relationship and learn better nodes representations for nodes from novel classes in GFS-NC task. GraphKD consists of two modules: balanced pretraining module and class-relation transferring module. Balanced pretraining can optimize network parameters to a suitable manifold for subsequent initialization. The class-relation transferring module leverages a knowledge distillation model, where a teacher model generates soft labels containing interclass relationships and then transfer them to the student model. The student model is optimized to fit both the soft labels and hard labels concurrently. This relationship information can help the student model better understand the similarities and differences between classes, thereby improving its classification performance. In addition, we employee information entropy to distinguish the samples locate at the boundary of a base class and novel class and then assign them larger weights in the student model to enhance its expressive capacity for novel nodes. Our experiments show that the proposed method outperforms state-of-the-art baselines on various few-shot node classification datasets. |
Keyword | Few-shot Learning (Fsl) Graph Neural Networks (Gnns) Knowledge Distillation Node Classification |
DOI | 10.1109/TCSS.2024.3382471 |
URL | View the original |
Indexed By | SCIE |
Language | 英語English |
WOS Research Area | Computer Science |
WOS Subject | Computer Science, Cybernetics ; Computer Science, Information Systems |
WOS ID | WOS:001324994700001 |
Publisher | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC, 445 HOES LANE, PISCATAWAY, NJ 08855-4141 |
Scopus ID | 2-s2.0-85190814277 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU) DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE |
Corresponding Author | Gong, Zhiguo |
Affiliation | State Key Laboratory of Internet of Things for Smart City and the Department of Computer and Information Science, University of Macau, Macao, China |
First Author Affilication | University of Macau |
Corresponding Author Affilication | University of Macau |
Recommended Citation GB/T 7714 | Wang, Jialong,Zhou, Mengting,Zhang, Shilong,et al. Generalized Few-Shot Node Classification With Graph Knowledge Distillation[J]. IEEE Transactions on Computational Social Systems, 2024. |
APA | Wang, Jialong., Zhou, Mengting., Zhang, Shilong., & Gong, Zhiguo (2024). Generalized Few-Shot Node Classification With Graph Knowledge Distillation. IEEE Transactions on Computational Social Systems. |
MLA | Wang, Jialong,et al."Generalized Few-Shot Node Classification With Graph Knowledge Distillation".IEEE Transactions on Computational Social Systems (2024). |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment