Abstract

In response to the background penetration problem of unsupervised style transfer algorithms in most cases, a Transformer style transfer network DualGGAN based on dual generators and fusion of relative position encoding is proposed. The network is trained using the least squares generative adversarial network, and the neural network is used as the image feature extractor to generate feature maps to obtain facial image features with attention weights from the feature maps, utilizing relative position encoding and mask loss to jointly constrain feature region style transfer. The experimental results show that the DualGGAN network effectively reduces artifact generation when implementing facial style transfer, maintains good background consistency, and has good generalization ability. Experiments have shown that the FID and KID indicators on the cat2dog and older2adult datasets are significantly improved compared to other algorithms.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call