UM  > Faculty of Science and Technology
Residential Collegefalse
Status已發表Published
Towards Demonstration-Aware Large Language Models for Machine Translation
Chen Li1; Meishan Zhang1; Xuebo Liu1; Zhaocong Li2; Derek F. Wong2; Min Zhang1
2024
Conference NameThe 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024)
Source PublicationFindings of the 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024)
Pages13868-13881
Conference Date11-16 August 2024
Conference PlaceBangkok
CountryThailand
PublisherAssociation for Computational Linguistics (ACL)
Abstract

Tuning-based large language models for machine translation (aka large translation model, LTM) have demonstrated significant performance in the field of machine translation. Despite their success, these models often face difficulties in leveraging demonstrations to further improve their performance. To tackle this challenge, we introduce a novel approach that integrates demonstration-aware training and inference strategies within the framework of tuning-based LTMs, hereby referred to as demonstration-aware LTMs. During training, we enrich the model's learning process by incorporating both sentence- and document-level demonstrations derived from its original training dataset. During inference, the model synergizes its own contextual translations with retrieved high-quality demonstrations, leading to more precise and contextually appropriate outputs. Empirical results reveal that our demonstration-aware LTM not only mitigates the negative impacts traditionally associated with demonstrations but also secures substantial improvements in translation accuracy, particularly in domain-specific and document-level translation tasks. Source code and scripts are freely available at https://github.com/ChenLi0620/Demo-Aware-LLM-MT.

DOI10.18653/v1/2024.findings-acl.824
Language英語English
Scopus ID2-s2.0-85205325742
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionFaculty of Science and Technology
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorXuebo Liu
Affiliation1.Institute of Computing and Intelligence, Harbin Institute of Technology, Shenzhen, China
2.NLP2CT Lab, Department of Computer and Information Science, University of Macau
Recommended Citation
GB/T 7714
Chen Li,Meishan Zhang,Xuebo Liu,et al. Towards Demonstration-Aware Large Language Models for Machine Translation[C]:Association for Computational Linguistics (ACL), 2024, 13868-13881.
APA Chen Li., Meishan Zhang., Xuebo Liu., Zhaocong Li., Derek F. Wong., & Min Zhang (2024). Towards Demonstration-Aware Large Language Models for Machine Translation. Findings of the 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024), 13868-13881.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Chen Li]'s Articles
[Meishan Zhang]'s Articles
[Xuebo Liu]'s Articles
Baidu academic
Similar articles in Baidu academic
[Chen Li]'s Articles
[Meishan Zhang]'s Articles
[Xuebo Liu]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Chen Li]'s Articles
[Meishan Zhang]'s Articles
[Xuebo Liu]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.