Abstract

<span lang="EN-US">Classification of the types of weld defects is one of the stages of evaluating radiographic images, which is an essential step in controlling the quality of welded joints in materials. By automating the weld defects classification based on deep learning and the CNN architecture, it is possible to overcome the limitations of visually or manually evaluating radiographic images. Good accuracy in classification models for weld defects requires the availability of sufficient datasets. In reality, however, the radiographic image dataset accessible to the public is limited and imbalanced between classes. Consequently, simple image cropping and augmentation techniques are implemented during the data preparation stage. To construct a weld defect classification model, we proposed to utilize the transfer learning method by employing a pre-trained CNN architecture as a feature extractor, including DenseNet201, InceptionV3, MobileNetV2, NASNetMobile, ResNet50V2, VGG16, VGG19, and Xception, which are linked to a simple classification model based on multilayer perceptron. The test results indicate that the three best classification models were obtained by using the DenseNet201 feature extractor with a test accuracy value of 100%, followed by ResNet50V2 and InceptionV3 with an accuracy of 99.17%. These outcomes are better compared to state-of-the-art classification models with a maximum of six classes of defects. The research findings may assist radiography experts in evaluating radiographic images more accurately and efficiently.</span>

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call