Residential College | false |
Status | 已發表Published |
Noise Stability Regularization for Improving BERT Fine-tuning | |
Hua, H.H.1; Li, X.J.2,3; Dou, D.J.2; Xu, C.Z.3; Luo, J.B.1 | |
2021-06-07 | |
Conference Name | 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021 |
Source Publication | NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference |
Pages | 3229 - 3241 |
Conference Date | JUN 06-11, 2021 |
Conference Place | Virtual, Online |
Publication Place | STROUDSBURG, PA 18360 USA |
Publisher | Association for Computational Linguistics (ACL) |
Abstract | Fine-tuning pre-trained language models such as BERT has become a common practice dominating leaderboards across various NLP tasks. Despite its recent success and wide adoption, this process is unstable when there are only a small number of training samples available. The brittleness of this process is often reflected by the sensitivity to random seeds. In this paper, we propose to tackle this problem based on the noise stability property of deep nets, which is investigated in recent literature (Arora et al., 2018; Sanyal et al., 2020). Specifically, we introduce a novel and effective regularization method to improve fine-tuning on NLP tasks, referred to as Layer-wise Noise Stability Regularization (LNSR). We extend the theories about adding noise to the input and prove that our method gives a stabler regularization effect. We provide supportive evidence by experimentally confirming that well-performing models show a low sensitivity to noise and fine-tuning with LNSR exhibits clearly higher generalizability and stability. Furthermore, our method also demonstrates advantages over other state-of-the-art algorithms including L2 - SP (Li et al., 2018), Mixout (Lee et al., 2020) and SMART (Jiang et al., 2020). |
Keyword | -- |
DOI | 10.48550/arXiv.2107.04835 |
URL | View the original |
Indexed By | CPCI-S ; CPCI-SSH |
Language | 英語English |
WOS Research Area | Computer Science ; Linguistics |
WOS Subject | Computer Science, Artificial Intelligence ; Linguistics |
WOS ID | WOS:000895685603030 |
The Source to Article | PB_Publication |
Scopus ID | 2-s2.0-85114436090 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE Faculty of Science and Technology |
Corresponding Author | Dou, D.J. |
Affiliation | 1.University of Rochester, Rochester, NY, USA 2.Big Data Lab, Baidu Research, Beijing, China 3.Department of Computer Science, University of Macau, Macau, China |
Recommended Citation GB/T 7714 | Hua, H.H.,Li, X.J.,Dou, D.J.,et al. Noise Stability Regularization for Improving BERT Fine-tuning[C], STROUDSBURG, PA 18360 USA:Association for Computational Linguistics (ACL), 2021, 3229 - 3241. |
APA | Hua, H.H.., Li, X.J.., Dou, D.J.., Xu, C.Z.., & Luo, J.B. (2021). Noise Stability Regularization for Improving BERT Fine-tuning. NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference, 3229 - 3241. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment