Residential Collegefalse
Status已發表Published
On the Complementarity between Pre-Training and Back-Translation for Neural Machine Translation
Liu, Xuebo1; Wang, Longyue2; Wong, Derek F.1; Ding, Liang3; Chao, Lidia S.1; Shi, Shuming2; Tu, Zhaopeng2
2021
Conference Name2021 Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021
Source PublicationFindings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021
Pages2900-2907
Conference Date7 November 2021through 11 November 2021
Conference PlacePunta Cana
CountryThe Dominican Republic
Author of SourceMoens M.-F., Huang X., Specia L., Yih S.W.-T.
PublisherAssociation for Computational Linguistics (ACL)
Abstract

Pre-training (PT) and back-translation (BT) are two simple and powerful methods to utilize monolingual data for improving the model performance of neural machine translation (NMT). This paper takes the first step to investigate the complementarity between PT and BT. We introduce two probing tasks for PT and BT respectively and find that PT mainly contributes to the encoder module while BT brings more benefits to the decoder. Experimental results show that PT and BT are nicely complementary to each other, establishing state-ofthe-art performances on the WMT16 EnglishRomanian and English-Russian benchmarks. Through extensive analyses on sentence originality and word frequency, we also demonstrate that combining Tagged BT with PT is more helpful to their complementarity, leading to better translation quality. Source code is freely available at https://github.com/ SunbowLiu/PTvsBT.

DOI10.48550/arXiv.2110.01811
URLView the original
Language英語English
Scopus ID2-s2.0-85115837324
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionDEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorLiu, Xuebo
Affiliation1.NLP2CT Lab, Department of Computer and Information Science, University of Macau, Macao
2.Tencent AI Lab, China
3.The University of Sydney, Australia
First Author AffilicationUniversity of Macau
Corresponding Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Liu, Xuebo,Wang, Longyue,Wong, Derek F.,et al. On the Complementarity between Pre-Training and Back-Translation for Neural Machine Translation[C]. Moens M.-F., Huang X., Specia L., Yih S.W.-T.:Association for Computational Linguistics (ACL), 2021, 2900-2907.
APA Liu, Xuebo., Wang, Longyue., Wong, Derek F.., Ding, Liang., Chao, Lidia S.., Shi, Shuming., & Tu, Zhaopeng (2021). On the Complementarity between Pre-Training and Back-Translation for Neural Machine Translation. Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021, 2900-2907.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Liu, Xuebo]'s Articles
[Wang, Longyue]'s Articles
[Wong, Derek F.]'s Articles
Baidu academic
Similar articles in Baidu academic
[Liu, Xuebo]'s Articles
[Wang, Longyue]'s Articles
[Wong, Derek F.]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Liu, Xuebo]'s Articles
[Wang, Longyue]'s Articles
[Wong, Derek F.]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.