UM  > Faculty of Arts and Humanities
Residential Collegefalse
Status已發表Published
FOCUS: Forging Originality through Contrastive Use in Self-Plagiarism for Language Models
Kaixin Lan1; Tao Fang1; Derek F. Wong1; Yabo Xu2; Lidia S. Chao1; Cecilia Guanfang Zhao3
2024
Conference NameThe 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024)
Source PublicationFindings of the 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024)
Pages14432-14447
Conference Date11-16 August 2024
Conference PlaceBangkok
CountryThailand
PublisherAssociation for Computational Linguistics (ACL)
Abstract

Pre-trained Language Models (PLMs) have shown impressive results in various Natural Language Generation (NLG) tasks, such as powering chatbots and generating stories. However, an ethical concern arises due to their potential to produce verbatim copies of paragraphs from their training data. This is problematic as PLMs are trained on corpora constructed by human authors. As such, there is a pressing need for research to promote the generation of original content by these models. In this study, we introduce a unique “self-plagiarism” contrastive decoding strategy, aimed at boosting the originality of text produced by PLMs. Our method entails modifying prompts in LLMs to develop an amateur model and a professional model. Specifically, the amateur model is urged to plagiarize using three plagiarism templates we have designed, while the professional model maintains its standard language model status. This strategy employs prompts to stimulate the model’s capacity to identify non-original candidate token combinations and subsequently impose penalties. The application of this strategy is integrated prior to the model’s final layer, ensuring smooth integration with most existing PLMs (T5, GPT, LLaMA) without necessitating further adjustments. Implementing our strategy, we observe a significant decline in non-original sequences comprised of more than three words in the academic AASC dataset and the story-based ROCStories dataset. 

DOI10.48550/arXiv.2406.00839
URLView the original
Language英語English
Scopus ID2-s2.0-85197182539
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionFaculty of Arts and Humanities
Faculty of Science and Technology
DEPARTMENT OF ENGLISH
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorDerek F. Wong
Affiliation1.NLP2CT Lab, Department of Computer and Information Science, University of Macau
2.Guangdong Hengqin DataStory Information Technology Ltd.
3.Department of English, Faculty of Arts and Humanities, University of Macau
First Author AffilicationUniversity of Macau
Corresponding Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Kaixin Lan,Tao Fang,Derek F. Wong,et al. FOCUS: Forging Originality through Contrastive Use in Self-Plagiarism for Language Models[C]:Association for Computational Linguistics (ACL), 2024, 14432-14447.
APA Kaixin Lan., Tao Fang., Derek F. Wong., Yabo Xu., Lidia S. Chao., & Cecilia Guanfang Zhao (2024). FOCUS: Forging Originality through Contrastive Use in Self-Plagiarism for Language Models. Findings of the 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024), 14432-14447.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Kaixin Lan]'s Articles
[Tao Fang]'s Articles
[Derek F. Wong]'s Articles
Baidu academic
Similar articles in Baidu academic
[Kaixin Lan]'s Articles
[Tao Fang]'s Articles
[Derek F. Wong]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Kaixin Lan]'s Articles
[Tao Fang]'s Articles
[Derek F. Wong]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.