Abstract

Human activity recognition aims to work out the activities performed by someone in a picture or video. Examples of actions are running, sitting, sleeping, and standing. Complex movement patterns and harmful occurrences like falling may be a part of these activities. The suggested ConvLSTM network can be created by successively combining fully connected layers, long immediate memory (LSTM) networks, and convolutional neural networks (CNN). The acquisition system will pre calculate skeleton coordinates using human detection and pose estimation from the image/video sequence. The ConvLSTM model builds new controlled features from the raw skeleton coordinates and their distinctive geometric and kinematic properties. Raw skeleton coordinates are utilized to generate geometric and kinematic properties supported by relative joint position values, joint differences, and their angular velocities. By utilizing a multi-player trained CNN-LSTM combination, novel spatiotemporal directed features can be obtained. The classification head with completely connected layers is then utilized. The suggested model was tested using the KinectHAR dataset, which consists of 130,000 samples with 81 attribute variables and was compiled using the Kinect (v2) sensor Experimental data is used to compare the performance of independent CNN and LSTM networks.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.