Abstract

Background: In recent years, human activity recognition (HAR) has been an active research topic due to its widespread application in various fields such as healthcare, sports, patient monitoring, etc. HAR approaches can be categorised as handcrafted feature methods (HCF) and deep learning methods (DL). HCF involves complex data pre-processing and manual feature extraction in which the models may be exposed to high bias and crucial implicit pattern loss. Hence, DL approaches are introduced due to their exceptional recognition performance. Convolutional Neural Network (CNN) extracts spatial features while preserving localisation. However, it hardly captures temporal features. Recurrent Neural Network (RNN) learns temporal features, but it is susceptible to gradient vanishing and suffers from short-term memory problems. Unlike RNN, Long-Short Term Memory network has a relatively longer-term dependency. However, it consumes higher computation and memory because it computes and stores partial results at each level. Methods: This work proposes a novel multiscale temporal convolutional network (MSTCN) based on the Inception model with a temporal convolutional architecture. Unlike HCF methods, MSTCN requires minimal pre-processing and no manual feature engineering. Further, multiple separable convolutions with different-sized kernels are used in MSTCN for multiscale feature extraction. Dilations are applied to each separable convolution to enlarge the receptive fields without increasing the model parameters. Moreover, residual connections are utilised to prevent information loss and gradient vanishing. These features enable MSTCN to possess a longer effective history while maintaining a relatively low in-network computation. Results: The performance of MSTCN is evaluated on UCI and WISDM datasets using subject independent protocol with no overlapping subjects between the training and testing sets. MSTCN achieves F1 scores of 0.9752 on UCI and 0.9470 on WISDM. Conclusion: The proposed MSTCN dominates the other state-of-the-art methods by acquiring high recognition accuracies without requiring any manual feature engineering.

Highlights

  • Human activity recognition (HAR) is extensively applied in various applications such as personal health monitoring,[1,2] geriatric patient monitoring,[3] ambient assisted living,[4] etc

  • The segmented data was fed into seven multiscale dilation (MSD) blocks in multiscale temporal convolutional network (MSTCN) for feature extraction

  • The proposed MSTCN showed extraordinary performances against the existing methods by achieving 97.46% accuracy on UCI and 95.20% on WISDM

Read more

Summary

Introduction

Human activity recognition (HAR) is extensively applied in various applications such as personal health monitoring,[1,2] geriatric patient monitoring,[3] ambient assisted living,[4] etc. Numerous HAR algorithms have been proposed, including handcrafted feature (HCF) methods[5,6,7] and deep learning (DL) methods.[8,9,10] HCF methods require complex data pre-processing and manual feature engineering. HCF involves complex data preprocessing and manual feature extraction in which the models may be exposed to high bias and crucial implicit pattern loss. Recurrent Neural Network (RNN) learns temporal features, but it is susceptible to gradient vanishing and suffers from short-term memory problems. Unlike HCF methods, MSTCN requires minimal pre-processing and no manual feature engineering. Residual connections are utilised to prevent information loss and gradient vanishing These features enable MSTCN to possess a longer effective history while maintaining a relatively low in-network computation. MSTCN achieves F1 scores of 0.9752 on UCI and 0.9470 on WISDM

Methods
Findings
Discussion
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.