Abstract

Even though a linguistics-free sequence to sequence model in neural machine translation (NMT) has certain capability of implicitly learning syntactic information of source sentences, this paper shows that source syntax can be explicitly incorporated into NMT effectively to provide further improvements. Specifically, we linearize parse trees of source sentences to obtain structural label sequences. On the basis, we propose three different sorts of encoders to incorporate source syntax into NMT: 1) Parallel RNN encoder that learns word and label annotation vectors parallelly; 2) Hierarchical RNN encoder that learns word and label annotation vectors in a two-level hierarchy; and 3) Mixed RNN encoder that stitchingly learns word and label annotation vectors over sequences where words and labels are mixed. Experimentation on Chinese-to-English translation demonstrates that all the three proposed syntactic encoders are able to improve translation accuracy. It is interesting to note that the simplest RNN encoder, i.e., Mixed RNN encoder yields the best performance with an significant improvement of 1.4 BLEU points. Moreover, an in-depth analysis from several perspectives is provided to reveal how source syntax benefits NMT.

Highlights

  • The sequence to sequence model in neural machine translation (NMT) has achieved certain success over the state-ofthe-art of statistical machine translation (SMT)NP1 VV NP2 input: ӳՂ ᦤԻಅ ಢ‫ اٴ‬ෛኞ ᱷᤈ ኩ᧗ Ӥ૱ ຫນ output: tokoyo stock exchange approves new listing bank reference: tokyo exchange approves shinsei bank 's application for listing (a)

  • We have investigated whether and how source syntax can explicitly help NMT to improve its translation accuracy

  • Experimentation on Chinese-to-English translation shows that all proposed models yield improvements over a state-ofthe-art baseline NMT system

Read more

Summary

Introduction

The sequence to sequence model (seq2seq) in neural machine translation (NMT) has achieved certain success over the state-ofthe-art of statistical machine translation (SMT). Shi et al (2016) show that the seq2seq model still fails to capture a lot of deep structural details, even though it is capable of learning certain implicit source syntax from sentence-aligned parallel corpus. It requires an additional parsing-task-specific training mechanism to recover the hidden syntax in NMT. In the absence of explicit linguistic knowledge, the seq2seq model in NMT tends to produce translations that fail to well respect syntax. Statistics on our development set show that one forth of Chinese noun phrases are translated into discontinuous phrases in English, indicating the substantial disrespect of syntax in NMT translation.. Experimentation on Chinese-to-English translation demonstrates that all proposed approaches are able to improve the translation accuracy

Attention-based NMT
NMT with Source Syntax
Syntax Representation
Iove ew3
RNN Encoders with Source Syntax
Comparison of RNN Encoders with Source Syntax
Experimental Settings
Experiment Results
Effects on Long Sentences
Analysis on Word Alignment
Analysis on Phrase Alignment
Analysis on Over Translation
Analysis on Rare Word Translation
Related Work
Conclusion

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.