In this study, we focus on human activity recognition, particularly aiming to distinguish the activity of praying (salat) from other daily activities. To achieve this goal, we have created a new dataset named HAR-P (Human activity recognition for Praying), which includes eight different activities: walking, running, sitting, standing, walking upstairs, walking downstairs, typing with a keyboard, and praying (salat). The HAR-P dataset was collected from 50 male individuals, who wore smartwatches on their dominant wrists. We compare the activity classification performance using three state-of-the-art algorithms from the literature: Long Short-Term Memory, Convolutional Long Short-Term Memory, and Convolutional Neural Network—Long Short-Term Memory. To assess the influence of sensors, data from accelerometer, gyroscope, linear acceleration sensor, and magnetic field sensor were utilized. The impact of individual sensor data as well as combinations thereof was investigated. The highest classification accuracy within single sensor groups, reaching 95.7%, was achieved using the accelerometer data with the Convolutional Long Short-Term Memory method. Combining two sensor groups resulted in an increase in accuracy of up to 9%. The highest accuracy of 96.4% was obtained by utilizing three sensor groups together with the Convolutional Neural Network—Long Short-Term Memory method. Furthermore, the evaluation of sensor and model performance was conducted using the stratified k-fold cross-validation method with 5-folds. These findings contribute significantly to evaluating the performance of sensor combinations and different algorithms in activity classification. This study may provide an effective foundation for the automatic recognition and tracking of human activities and offer an applicable model, particularly for the recognition of religious practices such as praying.
Read full abstract