Abstract

Efficient and differentiable image over-segmentation is key to superpixel-based research and applications but remains a challenging problem. The paper proposes a fully convolutional deep network, named OverSegNet, for image over-segmentation. OverSegNet consists of an encoder and a decoder, which are designed for the two core parts of over-segmentation, i.e., feature representation and pixel–superpixel association, respectively. To obtain edge-sensitive and noise-insusceptible feature representation, the encoder is endowed with rich over-segmentation-specific convolutional kernels via over-parametrization followed by task-driven neural network search (NAS). The decoder adopts a multi-scale convolutional structure with cross-large-scale connections, to achieve pixel–superpixel association in a coarse-to-fine feed-forward manner while eliminating accumulation errors. We conduct rich ablation studies to verify the effectiveness of the specially designed encoder and decoder. Experiments on both the BSDS500 dataset and NYUv2 dataset show that the proposed OverSegNet is fast, obtains state-of-the-art accuracy and has good generalization ability. Using semantic segmentation and disparity estimation as examples, we also verify the proposed OverSegNet in downstream applications.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.