Abstract

Deep-learning based generative models have achieved outstanding performance in various image processing tasks. This paper introduces a method to address the problem of extrapolating or outpainting visual context. When the input size is a small proportion of the output size, a limited amount of information is present to regenerate a semantically coherent image. This task is challenging because the missing region of the original image may include crucial semantic and spatial structural information, which is difficult to predict from the input. We propose a three-stage edge-guided coarse-to-fine generative network model, consisting of a contextual inference network, structural edge map generator and edge enhanced network, to synthesise semantically consistent output from small picture inputs. Our model adopts a gradual growth inference strategy in the contextual inference network so that the generated image can present a more coherent structure, and this result can support the structural edge map generator to generate a reasonable edge map in a large missing area. Combining the contextual inference network and structural edge map generator outputs enables the edge enhanced network to generate more convincing images. We evaluate our model using four public datasets: CelebA, Places2, Oxford Flower102and CUB200. Our experimental results demonstrate that the proposed image outpainting network can successfully regenerate high-quality images with a large missing region even when some structural features are lost in the input images.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call