Advancements in automation and artificial intelligence have significantly impacted accessibility for individuals with visual impairments, particularly in the realm of bus public transportation. Effective bus detection and bus point-of-view (POV) classification are crucial for enhancing the independence of visually impaired individuals. This study introduces the Improved-YOLOv10, a novel model designed to tackle challenges in bus identification and pov classification by integrating Coordinate Attention (CA) and Adaptive Kernel Convolution (AKConv) into the YOLOv10 framework. The Improved YOLOv10 advances the YOLOv10 architecture through the incorporation of CA, which enhances long-range dependency modeling and spatial awareness, and AKConv, which dynamically adjusts convolutional kernels for superior feature extraction. These enhancements aim to improve both detection accuracy and efficiency, essential for real-time applications in assistive technologies. Evaluation results demonstrate that the Improved-YOLOv10 offers significant improvements in detection performance, including better Accuracy, Precision and Recall compared to YOLOv10. The model also exhibits reduced computational complexity and storage requirements, highlighting its efficiency. While the classification results show some trade-offs, with slightly decreased overall F1 score, the complexity of Giga Floating Point Operations (GFLOPs), Parameters, and Weight/MB in the Improved-YOLOv10 remains advantageous for classification tasks. The model’s architectural improvements contribute to its robustness and efficiency, making it a suitable choice for real-time applications and assistive technologies.
Read full abstract