Abstract

The training method that uses predefined degradation methods to provide training data for deep convolutional neural networks suffers from overfitting and image distortion when performing face super-resolution reconstruction. In this paper, we propose a multi-task facial super-resolution reconstruction framework embedded in a degraded augmented GAN network. Compared with the commonly used method of generating low-quality face images through pixel interpolation, this paper inversely embeds a generative adversarial network composed of residual coding blocks in the network. Through the feature learning ability of the network, the feature fitting of the face images collected in the natural scene is performed to generate training data that can more fully represent the real noise in the natural scene, thereby improving the noise reduction and image reconstruction capabilities of the network. In order to take full advantage of the task-specific features, the semantic features of the dataset are extracted through a pre-trained facial organ semantic detection framework and the data is used to add face semantic information in the super-resolution reconstruction network. The experimental results show that the model proposed in this study can achieve super-resolution reconstruction of ultra-low-resolution face images simply and efficiently, and achieve super-resolution face images with lower visual quality than other advanced methods with lower model complexity. It solves the problem of single style and distortion of face reconstruction images caused by the training method of the data set obtained by the usual predefined degradation.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call