Abstract

Semantic segmentation is an important task in computer vision that aims to infer pixel-level semantic label information in images. Recently significant progresses have been made in deep neural networks-driven segmentation techniques, but these often require a large number of labels to supervise neural network training. Obtaining a sufficient number of labeled training data is challenging and sometimes impractical in real-world applications. This paper aims to study a semi-supervised semantic segmentation via image-to-image (I2I) translation technique. I2I is an emerging technique that maps an image from a domain to another domain. We uniquely treat the image semantic segmentation as an I2I translation task that infers semantic labels of objects (target domain) from input image (source domain) in a weakly supervised way. Particularly, we develop a two-pass strategy of I2I combining images with real and pseudo labels for semi-supervised model learning. The first pass uses unsupervised models to generate pseudo labels that combine with inputs to form pseudo-labeled samples. Since the pseudo-labeled images may undermine the quality of the model, they have to be specifically constrained in training by a noise correction framework to ensure good performance. Therefore, we boost the performance by incorporating both real and the pseudo labeled samples into the second pass to train a model based on a supervised architecture. Our goal is to bridge the gap between supervised and unsupervised learning for semantic object segmentation in practical. Extensive evaluations are conducted to demonstrate the efficiency and effectiveness of the proposed technique.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call