Residential Collegefalse
Status已發表Published
Improving Pretrained Language Model Fine-Tuning With Noise Stability Regularization
Hua, Hang1; Li, Xingjian2; Dou, Dejing3; Xu, Cheng Zhong4; Luo, Jiebo1
2023-11-30
Source PublicationIEEE Transactions on Neural Networks and Learning Systems
ISSN2162-237X
Pages1-15
Abstract

The advent of large-scale pretrained language models (PLMs) has contributed greatly to the progress in natural language processing (NLP). Despite its recent success and wide adoption, fine-tuning a PLM often suffers from overfitting, which leads to poor generalizability due to the extremely high complexity of the model and the limited training samples from downstream tasks. To address this problem, we propose a novel and effective fine-tuning framework, named layerwise noise stability regularization (LNSR). Specifically, our method perturbs the input of neural networks with the standard Gaussian or in-manifold noise in the representation space and regularizes each layer’s output of the language model. We provide theoretical and experimental analyses to prove the effectiveness of our method. The empirical results show that our proposed method outperforms several state-of-the-art algorithms, such as $\text{L}^2$ norm and start point (L2-SP), Mixout, FreeLB, and smoothness inducing adversarial regularization and Bregman proximal point optimization (SMART). In addition to evaluating the proposed method on relatively simple text classification tasks, similar to the prior works, we further evaluate the effectiveness of our method on more challenging question-answering (QA) tasks. These tasks present a higher level of difficulty, and they provide a larger amount of training examples for tuning a well-generalized model. Furthermore, the empirical results indicate that our proposed method can improve the ability of language models to domain generalization.

KeywordDomain Generalization Fine-tuning In-domain Generalization Pretrained Language Models (Plms) Regularization
DOI10.1109/TNNLS.2023.3330926
URLView the original
Indexed BySCIE
Language英語English
WOS Research AreaComputer Science ; Engineering
WOS SubjectComputer Science, Artificial Intelligence ; Computer Science, Hardware & Architecture ; Computer Science, Theory & Methods ; Engineering, Electrical & Electronic
WOS IDWOS:001112832100001
PublisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC, 445 HOES LANE, PISCATAWAY, NJ 08855-4141
Scopus ID2-s2.0-85180285311
Fulltext Access
Citation statistics
Document TypeJournal article
CollectionTHE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU)
Faculty of Science and Technology
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorDou, Dejing
Affiliation1.Department of Computer Science, University of Rochester, Rochester, NY, USA
2.Computational Biology Department, Carnegie Mellon University, Pittsburgh, PA, USA
3.BCG in Greater China, Beijing, China
4.Faculty of Science and Technology, State Key Laboratory of IOTSC, University of Macau, SAR, China
Recommended Citation
GB/T 7714
Hua, Hang,Li, Xingjian,Dou, Dejing,et al. Improving Pretrained Language Model Fine-Tuning With Noise Stability Regularization[J]. IEEE Transactions on Neural Networks and Learning Systems, 2023, 1-15.
APA Hua, Hang., Li, Xingjian., Dou, Dejing., Xu, Cheng Zhong., & Luo, Jiebo (2023). Improving Pretrained Language Model Fine-Tuning With Noise Stability Regularization. IEEE Transactions on Neural Networks and Learning Systems, 1-15.
MLA Hua, Hang,et al."Improving Pretrained Language Model Fine-Tuning With Noise Stability Regularization".IEEE Transactions on Neural Networks and Learning Systems (2023):1-15.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Hua, Hang]'s Articles
[Li, Xingjian]'s Articles
[Dou, Dejing]'s Articles
Baidu academic
Similar articles in Baidu academic
[Hua, Hang]'s Articles
[Li, Xingjian]'s Articles
[Dou, Dejing]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Hua, Hang]'s Articles
[Li, Xingjian]'s Articles
[Dou, Dejing]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.