Residential College | false |
Status | 已發表Published |
Towards Demonstration-Aware Large Language Models for Machine Translation | |
Chen Li1; Meishan Zhang1; Xuebo Liu1; Zhaocong Li2; Derek F. Wong2; Min Zhang1 | |
2024 | |
Conference Name | The 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024) |
Source Publication | Findings of the 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024) |
Pages | 13868-13881 |
Conference Date | 11-16 August 2024 |
Conference Place | Bangkok |
Country | Thailand |
Publisher | Association for Computational Linguistics (ACL) |
Abstract | Tuning-based large language models for machine translation (aka large translation model, LTM) have demonstrated significant performance in the field of machine translation. Despite their success, these models often face difficulties in leveraging demonstrations to further improve their performance. To tackle this challenge, we introduce a novel approach that integrates demonstration-aware training and inference strategies within the framework of tuning-based LTMs, hereby referred to as demonstration-aware LTMs. During training, we enrich the model's learning process by incorporating both sentence- and document-level demonstrations derived from its original training dataset. During inference, the model synergizes its own contextual translations with retrieved high-quality demonstrations, leading to more precise and contextually appropriate outputs. Empirical results reveal that our demonstration-aware LTM not only mitigates the negative impacts traditionally associated with demonstrations but also secures substantial improvements in translation accuracy, particularly in domain-specific and document-level translation tasks. Source code and scripts are freely available at https://github.com/ChenLi0620/Demo-Aware-LLM-MT. |
DOI | 10.18653/v1/2024.findings-acl.824 |
Language | 英語English |
Scopus ID | 2-s2.0-85205325742 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | Faculty of Science and Technology DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE |
Corresponding Author | Xuebo Liu |
Affiliation | 1.Institute of Computing and Intelligence, Harbin Institute of Technology, Shenzhen, China 2.NLP2CT Lab, Department of Computer and Information Science, University of Macau |
Recommended Citation GB/T 7714 | Chen Li,Meishan Zhang,Xuebo Liu,et al. Towards Demonstration-Aware Large Language Models for Machine Translation[C]:Association for Computational Linguistics (ACL), 2024, 13868-13881. |
APA | Chen Li., Meishan Zhang., Xuebo Liu., Zhaocong Li., Derek F. Wong., & Min Zhang (2024). Towards Demonstration-Aware Large Language Models for Machine Translation. Findings of the 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024), 13868-13881. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment