Abstract

This paper proposes the You Only Look Once (YOLO) dependency fusing attention network (DFAN) detection algorithm, improved based on the lightweight network YOLOv4-tiny. It combines the advantages of fast speed of traditional lightweight networks and high precision of traditional heavyweight networks, so it is very suitable for the real-time detection of high-altitude safety belts in embedded equipment. In response to the difficulty of extracting the features of an object with a low effective pixel ratio—which is an object with a low ratio of actual area to detection anchor area in the YOLOv4-tiny network—we make three major improvements to the baseline network: The first improvement is introducing the atrous spatial pyramid pooling network after CSPDarkNet-tiny extracts features. The second is to propose the DFAN, while the third is to introduce the path aggregation network (PANet) to replace the feature pyramid network (FPN) of the original network and fuse it with the DFAN. According to the experimental results in the high-altitude safety belt dataset, YOLO-DFAN improves the accuracy by 5.13% compared with the original network, and its detection speed meets the real-time demand. The algorithm also exhibits a good improvement on the Pascal voc07+12 dataset.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.