Abstract

A lightweight rotational object detection algorithm, R-YOLOv5, is proposed to address the limitations of traditional object detection algorithms that do not consider the diversity of vehicle scales in drone images and fail to obtain information on rotation angles. The proposed algorithm incorporated an angle prediction branch and introduced a circular smooth label (CSL) angle classification method to make it suitable for detection scenarios based on rotational boxes. A cascaded Swin Transformer block (STrB) is used to reduce computational complexity during feature fusion in the backbone network, further enhancing semantic information and global perception capabilities for small objects. A feature enhancement attention module (FEAM) is proposed to improve the utilization of detailed information through local feature self-supervision. An adaptive spatial feature fusion structure (ASFF) is introduced, which employs features extracted from different levels of the backbone network to perform multi-scale feature fusion. The experimental results show that the detection accuracy reaches 84.91% on the Drone-Vehicle dataset and 90.23% on the UCAS-AOD remote sensing dataset. The lightweight model has a parameter count of only 2.02 million and can achieve 82.6 FPS for high-resolution images, which is significantly better than existing lightweight models and more suitable for real-time detection of rotating vehicles in dense scenes, making it suitable for deployment on a large majority of embedded platforms.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call