Abstract

Unsupervised text style transfer is full of challenges due to the lack of parallel data and difficulties in content preservation. In this paper, we propose a novel neural approach to unsupervised text style transfer which we refer to as Cycle-consistent Adversarial autoEncoders (CAE) trained from non-parallel data. CAE consists of three essential components: (1) LSTM autoencoders that encode a text in one style into its latent representation and decode an encoded representation into its original text or a transferred representation into a style-transferred text, (2) adversarial style transfer networks that use an adversarially trained generator to transform a latent representation in one style into a representation in another style, and (3) a cycle-consistent constraint that enhances the capacity of the adversarial style transfer networks in content preservation. The entire CAE with these three components can be trained end-to-end. Extensive experiments and in-depth analyses on two widely-used public datasets consistently validate the effectiveness of proposed CAE in both style transfer and content preservation against several strong baselines in terms of four automatic evaluation metrics and human evaluation.

Highlights

  • Unsupervised text style transfer is to rewrite a text in one style into a text in another style while the content of the text remains the same as much as possible without using any parallel data

  • In order to solve the issues above, we propose a cycle-consistent adversarial autoencoders (CAE) for unsupervised text style transfer

  • The cycleconsistent constraint enables consistent Adversarial autoEncoders (CAE) to yield the best reverse perplexity (RPPL) as it palliates mode dropping in style transfer

Read more

Summary

Introduction

Unsupervised text style transfer is to rewrite a text in one style into a text in another style while the content of the text remains the same as much as possible without using any parallel data. We do not have large-scale style-to-style parallel data to train a text style transfer model in a supervised way. It is difficult to preserve the content of a text when its style is transferred. To obtain good content preservation for text style transfer, various disentanglement approaches (Shen et al, 2017; Hu et al, 2017; Fu et al, 2018; Sudhakar et al, 2019) are proposed to separate the content and style of a text in the latent space. Content-style disentanglement is not achievable as content and style typically interact with each other in texts in subtle ways (Lample et al, 2019)

Objectives
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call