Visually Impaired People (VIP) face significant challenges in their daily lives, relying on others or trained dogs for assistance when navigating outdoors. Researchers have developed the Smart Stick (SS) system as a more effective aid than traditional ones to address these challenges. Developing and utilizing the SS systems for VIP improves mobility, reliability, safety, and accessibility. These systems help users by identifying obstacles and hazards, keeping VIP safe and efficient. This paper presents the design and real-world implementation of an SS using an Arduino Nano microcontroller, GPS, GSM module, heart rate sensor, ultrasonic sensor, moisture sensor, vibration motor, and Buzzer. Based on sensor data, the SS can provide warning signals to VIP about the presence of obstacles and hazards around them. Several Machine Learning (ML) algorithms were used to improve the SS alert decision accuracy. Therefore, this paper used sensor data to train and test ten ML algorithms to find the most effective alert decision accuracy. Based on the ML algorithms, the alert decision, including the presence of obstacles, environmental conditions, and user health conditions, was examined using several performance metrics. Results showed that the AdaBoost, Gradient boosting, and Random Forest ML algorithms outperformed others and achieved an AUC and specificity of 100%, with 99.9% accuracy, F1-score, precision, recall, and MCC in the cross-validation phase. Integrating sensor data with ML algorithms revealed that the SS enables VIP to live independently and move safely without assistance.
Read full abstract