Abstract
The prevalent approach for unsupervised text style transfer is disentanglement between content and style. However, it is difficult to completely separate style information from the content. Other approaches allow the latent text representation to contain style and the target style to affect the generated output more than the latent representation does. In both approaches, however, it is impossible to adjust the strength of the style in the generated output. Moreover, those previous approaches typically perform both the sentence reconstruction and style control tasks in a single model, which complicates the overall architecture. In this paper, we address these issues by separating the model into a sentence reconstruction module and a style module. We use the Transformer-based autoencoder model for sentence reconstruction and the adaptive style embedding is learned directly in the style module. Because of this separation, each module can better focus on its own task. Moreover, we can vary the style strength of the generated sentence by changing the style of the embedding expression. Therefore, our approach not only controls the strength of the style, but also simplifies the model architecture. Experimental results show that our approach achieves better style transfer performance and content preservation than previous approaches.
Highlights
Text style transfer is the task of modifying a text with a specified style attribute
If the input is completely separated, the text style is transferred by substituting the style component with a new target; for this approach, a discriminator is used to separate the style from a latent representation z, a compressed representation of the input text (Fu et al, 2018; John et al, 2019)
Trade-offs are induced between the accuracy and bilingual evaluation understudy (BLEU) score
Summary
Text style transfer is the task of modifying a text with a specified style attribute. Text style transfer has been solved by a supervised method using a parallel dataset (Jhamtani et al, 2017) containing pairs of source and target sentences. Obtaining this parallel dataset that achieves a total one-to-one correspondence with a specified style is often not possible. The discriminator can be applied to the output text so that the output text reflects the given attribute (Shen et al, 2017; Hu et al, 2017; Zhao et al, 2018)
Published Version (Free)
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have