Residential Collegefalse
Status已發表Published
Noise Stability Regularization for Improving BERT Fine-tuning
Hua, H.H.1; Li, X.J.2,3; Dou, D.J.2; Xu, C.Z.3; Luo, J.B.1
2021-06-07
Conference Name2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021
Source PublicationNAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference
Pages3229 - 3241
Conference DateJUN 06-11, 2021
Conference PlaceVirtual, Online
Publication PlaceSTROUDSBURG, PA 18360 USA
PublisherAssociation for Computational Linguistics (ACL)
Abstract

Fine-tuning pre-trained language models such as BERT has become a common practice dominating leaderboards across various NLP tasks. Despite its recent success and wide adoption, this process is unstable when there are only a small number of training samples available. The brittleness of this process is often reflected by the sensitivity to random seeds. In this paper, we propose to tackle this problem based on the noise stability property of deep nets, which is investigated in recent literature (Arora et al., 2018; Sanyal et al., 2020). Specifically, we introduce a novel and effective regularization method to improve fine-tuning on NLP tasks, referred to as Layer-wise Noise Stability Regularization (LNSR). We extend the theories about adding noise to the input and prove that our method gives a stabler regularization effect. We provide supportive evidence by experimentally confirming that well-performing models show a low sensitivity to noise and fine-tuning with LNSR exhibits clearly higher generalizability and stability. Furthermore, our method also demonstrates advantages over other state-of-the-art algorithms including L2 - SP (Li et al., 2018), Mixout (Lee et al., 2020) and SMART (Jiang et al., 2020).

Keyword--
DOI10.48550/arXiv.2107.04835
URLView the original
Indexed ByCPCI-S ; CPCI-SSH
Language英語English
WOS Research AreaComputer Science ; Linguistics
WOS SubjectComputer Science, Artificial Intelligence ; Linguistics
WOS IDWOS:000895685603030
The Source to ArticlePB_Publication
Scopus ID2-s2.0-85114436090
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionDEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Faculty of Science and Technology
Corresponding AuthorDou, D.J.
Affiliation1.University of Rochester, Rochester, NY, USA
2.Big Data Lab, Baidu Research, Beijing, China
3.Department of Computer Science, University of Macau, Macau, China
Recommended Citation
GB/T 7714
Hua, H.H.,Li, X.J.,Dou, D.J.,et al. Noise Stability Regularization for Improving BERT Fine-tuning[C], STROUDSBURG, PA 18360 USA:Association for Computational Linguistics (ACL), 2021, 3229 - 3241.
APA Hua, H.H.., Li, X.J.., Dou, D.J.., Xu, C.Z.., & Luo, J.B. (2021). Noise Stability Regularization for Improving BERT Fine-tuning. NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference, 3229 - 3241.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Hua, H.H.]'s Articles
[Li, X.J.]'s Articles
[Dou, D.J.]'s Articles
Baidu academic
Similar articles in Baidu academic
[Hua, H.H.]'s Articles
[Li, X.J.]'s Articles
[Dou, D.J.]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Hua, H.H.]'s Articles
[Li, X.J.]'s Articles
[Dou, D.J.]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.