UM  > Faculty of Arts and Humanities  > DEPARTMENT OF PORTUGUESE
Residential Collegefalse
Status已發表Published
How Does Pretraining Improve Discourse-Aware Translation?
Zhihong Huang1; Longyue Wang2; Siyou Liu1; Derek F. Wong1
2023-08
Conference Name24th International Speech Communication Association, Interspeech 2023
Source PublicationProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
Volume2023-August
Pages3899 - 3903
Conference Date2023/08/20-2023/08/23
Conference PlaceDublin
CountryIreland
PublisherInternational Speech Communication Association
Abstract

Pretrained language models (PLMs) have produced substantial improvements in discourse-aware neural machine translation (NMT), for example, improved coherence in spoken language translation. However, the underlying reasons for their strong performance have not been well explained. To bridge this gap, we introduce a probing task to interpret the ability of PLMs to capture discourse relation knowledge. We validate three state-of-the-art PLMs across encoder-, decoder-, and encoder-decoder-based models. The analysis shows that (1) the ability of PLMs on discourse modelling varies from architecture and layer; (2) discourse elements in a text lead to different learning difficulties for PLMs. Besides, we investigate the effects of different PLMs on spoken language translation. Through experiments on IWSLT2017 Chinese-English dataset, we empirically reveal that NMT models initialized from different layers of PLMs exhibit the same trends with the probing task. Our findings are instructive to understand how and when discourse knowledge in PLMs should work for downstream tasks. © 2023 International Speech Communication Association. All rights reserved.

KeywordDiscourse Linguistic Probing Machine Translation Pretrained Language Models Spoken Language
DOI10.21437/Interspeech.2023-1068
URLView the original
Language英語English
Scopus ID2-s2.0-85171534338
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionDEPARTMENT OF PORTUGUESE
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorDerek F. Wong
Affiliation1.University of Macau
2.Tencent AI Lab
First Author AffilicationUniversity of Macau
Corresponding Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Zhihong Huang,Longyue Wang,Siyou Liu,et al. How Does Pretraining Improve Discourse-Aware Translation?[C]:International Speech Communication Association, 2023, 3899 - 3903.
APA Zhihong Huang., Longyue Wang., Siyou Liu., & Derek F. Wong (2023). How Does Pretraining Improve Discourse-Aware Translation?. Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, 2023-August, 3899 - 3903.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Zhihong Huang]'s Articles
[Longyue Wang]'s Articles
[Siyou Liu]'s Articles
Baidu academic
Similar articles in Baidu academic
[Zhihong Huang]'s Articles
[Longyue Wang]'s Articles
[Siyou Liu]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Zhihong Huang]'s Articles
[Longyue Wang]'s Articles
[Siyou Liu]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.