Abstract

This study uses an end-to-end encoder-decoder structure to build a machine translation model, allowing the machine to automatically learn features and transform the corpus data into distributed representations. The word vector uses a neural network to achieve direct mapping. Research on constructing neural machine translation models for different neural network structures. Based on the translation model of the LSTM network, the gate valve mechanism reduces the gradient attenuation and improves the ability to process long-distance sequences. Based on the GRU network structure, the simplified processing is performed on the basis of it, which reduces the training complexity and achieves good performance. Aiming at the problem that some source language sequences of any length in the encoder are encoded into fixed-dimensional background vectors, the attention mechanism is introduced to dynamically adjust the degree of influence of the source language context on the target language sequence, and the translation model’s ability to deal with long-distance dependencies is improved. In order to better reflect the context information, this study further proposes a machine translation model based on two-way GRU and compares and analyzes multiple translation models to verify the effectiveness of the model’s performance improvement.

Highlights

  • Machine translation usually combines natural language processing and artificial intelligence

  • Conclusions is article mainly focuses on Chinese-English machine translation tasks. e encoder and decoder parts are constructed by the neural network, and the word vector method of distributed representation is adopted, source language and target language sequence

  • Different translation models are constructed for comparison and analysis of three kinds of neural networks: recurrent neural network, long-term short-term memory, and gated recurrent unit. e attention mechanism is added to improve the translation model’s processing of long-distance dependence

Read more

Summary

Introduction

Machine translation usually combines natural language processing and artificial intelligence. Compared with the previous model, statistical machine translation learns the conversion rules from the corpus, no longer needs to provide the language rules actively, and solves the bottleneck problem of knowledge acquisition. E rule-based machine translation method highly relies on the rules of language It has a certain degree of versatility, the cost of obtaining the rules is relatively high. E well-known professor Bengio in the field of deep learning proposed a language model based on the neural network in 2003, which effectively alleviated the problem of data sparseness through distributed representation. 3. Neural Network Translation Model Incorporating Attention e rule-based machine translation method highly relies on the rules of language. Neural Network Translation Model Incorporating Attention e rule-based machine translation method highly relies on the rules of language Statistical machine translation has achieved good results, but there are still many problems

Get the current descent distance
Mj hj hj
Source Language Series
Language Series
Niutrans RNN LSTM
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call