Abstract

Neural machine translation has significantly pushed forward the quality of the field. However, there are remaining big issues with the output translations and one of them is fairness. Neural models are trained on large text corpora which contain biases and stereotypes. As a consequence, models inherit these social biases. Recent methods have shown results in reducing gender bias in other natural language processing tools such as word embeddings. We take advantage of the fact that word embeddings are used in neural machine translation to propose a method to equalize gender biases in neural machine translation using these representations. Specifically, we propose, experiment and analyze the integration of two debiasing techniques over GloVe embeddings in the Transformer translation architecture. We evaluate our proposed system on the WMT English-Spanish benchmark task, showing gains up to one BLEU point. As for the gender bias evaluation, we generate a test set of occupations and we show that our proposed system learns to equalize existing biases from the baseline system.

Highlights

  • Language is one of the most interesting and complex skills used in our daily life, and may even be taken for granted on our ability to communicate

  • Natural language processing (NLP) is a subfield of artificial intelligence that focuses on making natural languages understandable to computers

  • The translation between different natural languages is a task for Machine Translation (MT)

Read more

Summary

Introduction

Language is one of the most interesting and complex skills used in our daily life, and may even be taken for granted on our ability to communicate. Neural MT has shown significant improvements on performance using deep learning techniques, which are algorithms that learn abstractions from data In recent years, these deep learning techniques have shown promising results in narrowing the gap between human-like performance with sequence-to-sequence learning approaches in a variety of tasks (Sutskever et al, 2014), improvements in combination of approaches such as attention (Bahdanau et al, 2014) and translation systems algorithms like the Transformer (Vaswani et al, 2017). The Transformer (Vaswani et al, 2017) is a deep learning architecture based on self-attention, which has shown better performance over previous systems. It is more efficient in using computational resources and has higher training speed than previous recurrent (Sutskever et al, 2014; Bahdanau et al, 2014) and convolutional models (Gehring et al, 2017). The encoder reads an input sentence to generate a representation which is later used by a decoder to produce a sentence output word by word

Objectives
Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call