Abstract

As the typical distant language pair, Chinese and Vietnamese vary widely in syntactic structure, which significantly influences the performance of Chinese-Vietnamese machine translation. To address this problem, we present a simple approach with a pre-reordering model for closing syntactic gaps of the Chinese-Vietnamese language pair. Specifically, we first propose an algorithm for recognizing the modifier inverse, one of the most representative syntactic different in Chinese-Vietnamese language pair. Then we pre-train a pre-reordering model based on the former recognition algorithm and incorporate it into the attention-based translation framework for syntactic different reordering. We conduct empirical studies on Chinese-Vietnamese neural machine translation task, the results show that our approach achieves average improvement of 2.75 BLEU points in translation quality over the baseline model. In addition, the translation fluency can be significantly improved by over 2.44 RIBES points.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call