Abstract

Document-level neural machine translation has yielded attractive improvements. However, majority of existing methods roughly use all context sentences in a fixed scope. They neglect the fact that different source sentences need different sizes of context. To address this problem, we propose an effective approach to select dynamic context so that the document-level translation model can utilize the more useful selected context sentences to produce better translations. Specifically, we introduce a selection module that is independent of the translation module to score each candidate context sentence. Then, we propose two strategies to explicitly select a variable number of context sentences and feed them into the translation module. We train the two modules end-to-end via reinforcement learning. A novel reward is proposed to encourage the selection and utilization of dynamic context sentences. Experiments demonstrate that our approach can select adaptive context sentences for different source sentences, and significantly improves the performance of document-level translation methods.

Highlights

  • Neural machine translation (NMT) has achieved great progress in recent years (Cho et al, 2014; Bahdanau et al, 2015; Luong et al, 2015; Vaswani et al, 2017), when fed an entire document, standard neural machine translation (NMT) systems translate sentences in isolation without considering the cross-sentence dependencies

  • We mainly focus on a general scenario, where document-level neural machine translation (DocNMT) translates sentences with the online source-side context sentences

  • Experiments show that our approach can significantly improve the performance of DocNMT models with the selected dynamic context sentences

Read more

Summary

Introduction

Neural machine translation (NMT) has achieved great progress in recent years (Cho et al, 2014; Bahdanau et al, 2015; Luong et al, 2015; Vaswani et al, 2017), when fed an entire document, standard NMT systems translate sentences in isolation without considering the cross-sentence dependencies. Document-level neural machine translation (DocNMT) methods are proposed to utilize source-side or target-side intersentence contextual information to improve translation quality over sentences in a document (Jean et al, 2017; Wang et al, 2017; Tiedemann and Scherrer, 2017; Tu et al, 2018; Kuang et al, 2018; Junczys-Dowmunt, 2019; Ma et al, 2020). There is still an issue that has received less attention: which context sentences should be used when translating a source sentence?. We conduct an experiment to verify an intuition: the translation of different source sentences requires different context. We obtain dynamic context sentences that achieve the best BLEU scores by traversing all the context combinations for each source sentence. Experiments indicate that only the limited context sentences are really useful, and they change with source sentences

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call