Automated ultrasonic testing (AUT) is a nondestructive testing (NDT) method widely employed in industries that hold substantial economic importance. To ensure accurate inspections of exclusive AUT data, expert operators invest considerable effort and time. While artificial intelligence (AI)-assisted tools, utilizing deep learning models trained on extensive in-laboratory B-scan images, whether they are augmented or synthetically generated, have demonstrated promising performance for automated ultrasonic interpretation, ongoing efforts are needed to enhance their accuracy and applicability. This is possible through the evaluation of their performance with experimental ultrasonic data. In this study, we introduced a real-world ultrasonic B-scan image dataset generated from proprietary recorded AUT data during industrial automated girth weld inspection in oil and gas pipelines. The goal of inspection in our dataset was detecting a common type of defect called lack of fusion (LOF). We experimentally evaluated deep learning models for automatic weld defect detection using this dataset. Our assessment covers the baseline performance of state-of-the-art (SOTA) models, including transformer-based models (DETR and Deformable DETR) and YOLOv8. Their flaw detection performance in ultrasonic B-scan images has not been reported before. The results show that, without heavy augmentations or architecture customization, YOLOv8 outperforms the other models with an F1 score of 0.814 on our test set.
Read full abstract