Targeting the potential risk of side-vehicle collisions when the visually impaired crosses roads, this study proposed a side collision detection model, including monocular distance estimation, multimodal real-world location estimation, future location prediction and collision warning strategies tailored for visually impaired pedestrians. The proposed model employs YOLOv8 and DeepSort for vehicle detection and tracking, utilizing shallow neural networks for distance estimation based on image information and vehicle position data. Predicted vehicle distances are combined with magnetic field sensor and GPS data to compute and store real-world vehicle locations, and these location data will be used for linear regression to forecast future locations. A warning strategy is then implemented to alert users. Experimental validation shows that the monocular distance estimation network has an Absolute Relative Error of 0.043 and an ALE (Average Localization Error) of 1.249m. In real-world location estimation, the view angle ALE is 0.019, and the location ALE is 1.778m. Regarding location prediction, the accuracy in distinguishing stationary and moving vehicles reaches 0.962, and the predicted curve, based on ground truth and predicted locations, exhibits good alignment. The proposed warning strategy, evaluated on Kitti Tracking Dataset and a self-created dataset, accurately detects the majority of potential collision risks.
Read full abstract