UM  > Faculty of Science and Technology
Residential Collegefalse
Status已發表Published
Overcoming Catastrophic Forgetting for Fine-Tuning Pre-trained GANs
Zhang, Zeren1; Li, Xingjian2; Hong, Tao1; Wang, Tianyang3; Ma, Jinwen1; Xiong, Haoyi2; Xu, Cheng Zhong4
2023-09-18
Conference NameEuropean Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ECML PKDD 2023
Source PublicationLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume14173 LNAI
Pages293-308
Conference Date2023/09/18-2023/09/22
Conference PlaceTurin
Abstract

The great transferability of DNNs has induced a popular paradigm of “pre-training & fine-tuning”, by which a data-scarce task can be performed much more easily. However, compared to the existing efforts made in the context of supervised transfer learning, fewer explorations have been made on effectively fine-tuning pre-trained Generative Adversarial Networks (GANs). As reported in recent empirical studies, fine-tuning GANs faces the similar challenge of catastrophic forgetting as in supervised transfer learning. This causes a severe capacity loss of the pre-trained model when adapting it to downstream datasets. While most existing approaches suggest to directly interfere parameter updating, this paper introduces novel schemes from another perspective, i.e. inputs and features, thus essentially focuses on data aspect. Firstly, we adopt a trust-region method to smooth the adaptation dynamics by progressively adjusting input distributions, aiming to avoid dramatic parameter changes, especially when the pre-trained GAN has no information of target data. Secondly, we aim to avoid the loss of the diversity of the generated results of the fine-tuned GAN. This is achieved by explicitly encouraging generated images to encompass diversified spectral components in their deep features. We theoretically study the rationale of the proposed schemes, and conduct extensive experiments on popular transfer learning benchmarks to demonstrate the superiority of the schemes. The code and corresponding supplemental materials are available at https://github.com/zezeze97/Transfer-Pretrained-Gan.

KeywordGenerative Adversarial Networks Transfer Learning
DOI10.1007/978-3-031-43424-2_18
URLView the original
Indexed ByCPCI-S
Language英語English
WOS Research AreaComputer Science
WOS SubjectComputer Science, Artificial Intelligence ; Computer Science, Theory & Methods
WOS IDWOS:001156142300018
Scopus ID2-s2.0-85179624418
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionFaculty of Science and Technology
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorMa, Jinwen
Affiliation1.School of Mathematical Sciences, Peking University, Beijing, 100871, China
2.Baidu Inc., Beijing, China
3.University of Alabama at Birmingham, Birmingham, 35294, United States
4.University of Macau, Macao
Recommended Citation
GB/T 7714
Zhang, Zeren,Li, Xingjian,Hong, Tao,et al. Overcoming Catastrophic Forgetting for Fine-Tuning Pre-trained GANs[C], 2023, 293-308.
APA Zhang, Zeren., Li, Xingjian., Hong, Tao., Wang, Tianyang., Ma, Jinwen., Xiong, Haoyi., & Xu, Cheng Zhong (2023). Overcoming Catastrophic Forgetting for Fine-Tuning Pre-trained GANs. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 14173 LNAI, 293-308.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Zhang, Zeren]'s Articles
[Li, Xingjian]'s Articles
[Hong, Tao]'s Articles
Baidu academic
Similar articles in Baidu academic
[Zhang, Zeren]'s Articles
[Li, Xingjian]'s Articles
[Hong, Tao]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Zhang, Zeren]'s Articles
[Li, Xingjian]'s Articles
[Hong, Tao]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.