Residential College | false |
Status | 已發表Published |
On the Complementarity between Pre-Training and Back-Translation for Neural Machine Translation | |
Liu, Xuebo1; Wang, Longyue2; Wong, Derek F.1; Ding, Liang3; Chao, Lidia S.1; Shi, Shuming2; Tu, Zhaopeng2 | |
2021 | |
Conference Name | 2021 Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021 |
Source Publication | Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021 |
Pages | 2900-2907 |
Conference Date | 7 November 2021through 11 November 2021 |
Conference Place | Punta Cana |
Country | The Dominican Republic |
Author of Source | Moens M.-F., Huang X., Specia L., Yih S.W.-T. |
Publisher | Association for Computational Linguistics (ACL) |
Abstract | Pre-training (PT) and back-translation (BT) are two simple and powerful methods to utilize monolingual data for improving the model performance of neural machine translation (NMT). This paper takes the first step to investigate the complementarity between PT and BT. We introduce two probing tasks for PT and BT respectively and find that PT mainly contributes to the encoder module while BT brings more benefits to the decoder. Experimental results show that PT and BT are nicely complementary to each other, establishing state-ofthe-art performances on the WMT16 EnglishRomanian and English-Russian benchmarks. Through extensive analyses on sentence originality and word frequency, we also demonstrate that combining Tagged BT with PT is more helpful to their complementarity, leading to better translation quality. Source code is freely available at https://github.com/ SunbowLiu/PTvsBT. |
DOI | 10.48550/arXiv.2110.01811 |
URL | View the original |
Language | 英語English |
Scopus ID | 2-s2.0-85115837324 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE |
Corresponding Author | Liu, Xuebo |
Affiliation | 1.NLP2CT Lab, Department of Computer and Information Science, University of Macau, Macao 2.Tencent AI Lab, China 3.The University of Sydney, Australia |
First Author Affilication | University of Macau |
Corresponding Author Affilication | University of Macau |
Recommended Citation GB/T 7714 | Liu, Xuebo,Wang, Longyue,Wong, Derek F.,et al. On the Complementarity between Pre-Training and Back-Translation for Neural Machine Translation[C]. Moens M.-F., Huang X., Specia L., Yih S.W.-T.:Association for Computational Linguistics (ACL), 2021, 2900-2907. |
APA | Liu, Xuebo., Wang, Longyue., Wong, Derek F.., Ding, Liang., Chao, Lidia S.., Shi, Shuming., & Tu, Zhaopeng (2021). On the Complementarity between Pre-Training and Back-Translation for Neural Machine Translation. Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021, 2900-2907. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment