Abstract
Though tremendous strides have been made in face detection, face detection remains a challenging problem due to scale variance. In this paper, we propose a smoothed attention network for performing scale-invariant face detection by taking advantage of feature fusion and context enhancement, which is dubbed SANet. To reduce the noise in the fused features at different levels, an Attention-guided Feature Fusion Module (AFFM) is designed. In addition, an exhaustive analysis of the role of attention mechanism on performance is conducted, which considers channel-wise, spatial-wise attentions and their combinations. To enrich the contextual information by using dilated convolution and avoid the gridding artifacts problem produced by dilated convolution, we propose a Smoothed Context Enhancement Module (SCEM). Our method achieves state-of-the-art results on the UFDD dataset and comparable performance on the WIDER FACE, FDDB, PASCAL Faces, and AFW datasets.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: IEEE Transactions on Biometrics, Behavior, and Identity Science
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.