• Image fusion based on deflectometry principle can enhance the contrast between defective and non-defective regions. • A benchmark dataset of specular surface defects was constructed for algorithms evaluation and comparison. • An end-to-end attention-based network framework for automatic surface inspection is proposed. • Proposed framework can maintain performance dealing with extremely unbalanced pixel classes. • Quantitative and qualitative tests show superior performance compared to the state-of-art methods. Automated defect detection on specular vehicle surface with limited features (of up to 0.7 mm in diameter or width) and extremely unbalanced pixel classes is a still challenge of product quality control in automotive industry. The traditional defect inspection on specular surface is usually performed by inspectors, which is subjective, unstable and unquantified. Also, due to the limited features of isolated defect regions and hand-crafted feature extraction models may not be able to coordinated with each other, it is difficult for traditional methods to achieve comparable learning performance with the deep network. To alleviate these problems, a novel end-to-end attention-based fully convolutional neural network framework -DeepInspection is proposed for automated defect inspection on specular surface. Specifically, a sequence fusion algorithm through the principle of deflectometry is introduced to enhance the contrast between defective regions (pixels with lower intensity) and non-defective regions (pixels with higher intensity). Then, an attention-based deep convolutional neural network (DCNN) with Atrous Spatial Pyramid Pooling (ASPP) was proposed to capture local-to-global feature presentation from pre-processed fused images. To verify the robustness and effectiveness of the proposed method, a benchmark dataset called DeepInspection160 with 160 manually labeled images is established. Although the defective pixels only account for 0.561% in the DeepInspection160 dataset, the proposed DeepInspection framework still surpasses several state-of-the-art specular surface inspection methods which achieves F1 score over 0.7513 (pixel level) and 0.8055 (individual connected components) on the proposed challenging dataset.
Read full abstract