Abstract
With the rapid development of wearable devices and technologies, hand segmentation remains a less explored direction in egocentric vision, which is very important for activity recognition, rehabilitation, robot self-learning, etc. To overcome the high cost of auxiliary equipment and pixel-level annotations, we present an un-supervised hand segmentation method for egocentric images. Firstly, a fully convolutional neural network (FCN) is pre-trained in source dataset containing pixel-level annotations. Then, in target dataset without labels, the network is re-trained with optimized masks, which are produced by modified local and global consistency learning (LLGC) based on pre-segmentation and superpixel features. Finally, hand segmentation is realized in an alternative way. Furthermore, to balance segmentation accuracy and the cost on labeling, we propose a new semi-supervised image segmentation framework with three sub-nets based on the optimized noisy masks and a small number of clean labeled data. Experimental results in two target datasets indicate that the proposed methods could achieve better performance than other methods.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.