Abstract

While many image colorization algorithms have recently shown the capability of producing plausible color versions from gray-scale photographs, they still suffer from limited semantic understanding. To address this shortcoming, we propose to exploit pixelated object semantics to guide image colorization. The rationale is that human beings perceive and distinguish colors based on the semantic categories of objects. Starting from an autoregressive model, we generate image color distributions, from which diverse colored results are sampled. We propose two ways to incorporate object semantics into the colorization model: through a pixelated semantic embedding and a pixelated semantic generator. Specifically, the proposed network includes two branches. One branch learns what the object is, while the other branch learns the object colors. The network jointly optimizes a color embedding loss, a semantic segmentation loss and a color generation loss, in an end-to-end fashion. Experiments on Pascal VOC2012 and COCO-stuff reveal that our network, when trained with semantic segmentation labels, produces more realistic and finer results compared to the colorization state-of-the-art.

Highlights

  • Color has been at the center stage of computer vision for decades (e.g., Swain and Ballard 1991; Comaniciu and Meer 1997; Pérez et al 2002; Khan et al 2009; van de Sande et al 2010; Lou et al 2015; Vondrick et al 2018)

  • We study the relationship between colorization and semantic segmentation

  • To arrive at image colorization with pixelated semantics, we start from an autoregressive model

Read more

Summary

Introduction

Color has been at the center stage of computer vision for decades (e.g., Swain and Ballard 1991; Comaniciu and Meer 1997; Pérez et al 2002; Khan et al 2009; van de Sande et al 2010; Lou et al 2015; Vondrick et al 2018). Many vision challenges, including object detection and visual tracking, benefit from color (Khan et al 2009, 2012; Danelljan et al 2014; Vondrick et al 2018).

Objectives
Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call