Detection of aerial activities, including drones and birds, has practical implications for automating bird surveys and developing radar systems for aerial object collision detection. Convolutional neural networks (CNNs) have been extensively utilized for image recognition and classification tasks, albeit prior research predominantly focuses on single-class 'drone' classification. However, a gap persists in achieving high accuracy for multi-class classification. To address the limitations of traditional CNNs, such as vanishing gradients and the necessity for numerous layers, this study introduces a novel model termed "MobVGG.” This model combines the architectures of MobileNetV2 and VGG16 to accurately classify images as either 'bird' or 'drone'. The dataset comprises 4212 images for each category of 'bird' and 'drone'. The stringent methodology was applied for dataset preparation and model training to ensure the reliability of the findings. Comparative analysis with previous research demonstrates that the proposed MobVGG model, trained on both 'bird' and 'drone' images, achieves superior accuracy (96 %) compared to benchmark studies. Our paper targets researchers and graduate students as its primary audience.
Read full abstract