Abstract

Blur detection is aimed at estimating the probability of each pixel being blurred or non-blurred in an image affected by motion or defocus blur. This task has gained considerable attention due to its promising application fields in computer vision. Accurate differentiation of anomalous regions (including the sharp but homogeneous regions and pseudo-sharp backgrounds) and motion-blurred regions are main challenges in blur detection, in which both conventional and recently developed blur detection methods have limited performance and low time efficiency. To address these issues, this paper develops an accurate and fast blur detection method for both motion and defocus blur using a new end-to-end deep neural network. First, a novel multi-input multi-loss encoder-decoder network (M-shaped) is proposed to learn rich hierarchical representations related to blur. Then, to resolve the problem shows that blur degree is susceptible to scales, we construct a pyramid ensemble model (PM-Net) consisting of different scales of M-shaped subnets and a unified fusion layer. The experiments demonstrate that the proposed PM-Net can accurately handle those challenging scenarios with anomalous regions for both defocus and motion blur. Our method performs better than previous state-of-the-art methods. It achieves the $F_{1}$ -score of 0.893 for only defocus blur and 0.884 for joint motion and defocus blur, both of which significantly surpass previous methods on the benchmark BDD dataset. We also test our PM-Net on another public CDD dataset composed of challenging defocused images. The proposed method also outperforms other published methods with an $F_{1}$ -score of 0.885. In addition, our proposed method is hundreds of times faster (millisecond) than other state-of-the-art methods (second). Moreover, our experiments also demonstrate that the PM-Net is robust to noise and has a good generalization property.

Highlights

  • Blur is a ubiquitous degradation phenomenon in images captured using photography systems

  • To resolve the scale ambiguity and handle challenging scenarios more effectively, we further propose a pyramid ensemble network consisting of different scales of M-shaped subnets and a unified fusion layer, named PYRAMID M-SHAPED NETWORK (PM-Net)

  • To resolve the scale ambiguity problem and handle challenging scenarios more effectively, we propose an ensemble model based on the ensemble learning theory, i.e. the pyramid M-shaped network (PM-Net)

Read more

Summary

INTRODUCTION

Blur is a ubiquitous degradation phenomenon in images captured using photography systems. Countermeasures to the scale ambiguity problem in those ConvNet-based methods are not effective enough yet, which weightily limits the model performance To address these issues, we seek to develop a fast and accurate detection method for both motion and defocus blur using a new end-to-end deep network from raw image directly to final result. To resolve the scale ambiguity and handle challenging scenarios more effectively, we further propose a pyramid ensemble network consisting of different scales of M-shaped subnets and a unified fusion layer, named PM-Net. The uniqueness of our algorithm is: (1) a novel M-shaped architecture equipped simultaneously with multi-scale input and multi-level loss; (2) a pyramid ensemble network capable of resolving scale ambiguity and accurately detecting homogeneous and pseudo-sharp anomalies (Fig. 2(e)) for both motion and defocus blur; (3) fast detection speed (millisecond-level), which is hundreds of times faster than other state-of-the-art methods (second-level)

RELATED WORKS
METHODS
EXPERIMENTS
Findings
CONCLUSION

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.