Abstract

In the field of computer vision, super-resolution reconstruction techniques based on deep learning have undergone considerable advancement; however, certain limitations remain, such as insufficient feature extraction and blurred image generation. To address these problems, we propose an image super-resolution reconstruction model based on a generative adversarial network. First, we employ a dual network structure in the generator network to solve the problem of insufficient feature extraction. The dual network structure is divided into an upsample subnetwork and a refinement subnetwork, which upsample and optimize a low-resolution image, respectively. In a scene with large upscaling factors, this structure can reduce the negative effect of noise and enhance the utilization of high-frequency details, thereby generating high-quality reconstruction results. Second, to generate sharper super-resolution images, we use the perceptual loss, which exhibits a fast convergence and excellent visual effect, to guide the generator network training. We apply the ResNeXt-50- $32\times 4\text{d}$ network, which has few parameters and a large depth, to calculate the loss to obtain a reconstructed super-resolution image that is highly realistic. Finally, we introduce the Wasserstein distance into the discriminator network to enhance the discrimination ability and stability of the model. Specifically, this distance is employed to eliminate the activation function in the last layer of the network and avoid the use of the logarithm in calculating the loss function. Extensive experiments on the DIV2K, Set5, Set14, and BSD100 datasets demonstrate the effectiveness of the proposed model.

Highlights

  • With the development of information technology and progress in the digital age, the amount of all types of information is rapidly increasing

  • The generator network consists of an upsample subnetwork and refinement subnetwork, which are used jointly to solve the problems of SR and optimization, to ensure that the network can extract a larger amount of image feature information

  • Compared with the images reconstructed using the bicubic and super-resolution convolutional neural network (SRCNN) methods, the image reconstructed using the proposed model exhibits a higher quality, richer texture details, better color matching with the original image, and better visual effects. These findings demonstrate the effectiveness of the proposed model, which can recover more high-frequency details and generate clear SR images

Read more

Summary

Introduction

With the development of information technology and progress in the digital age, the amount of all types of information is rapidly increasing. In the field of image processing, image super-resolution (SR) reconstruction is a fundamental and critical issue, as a key step in many image related applications, including object tracking [1], object detection [2], semantic segmentation [3], image annotation [4], image inpainting [5], and image classification [6]. In these applications, a higher resolution of the image corresponds to more satisfactory results

Objectives
Methods
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call