Abstract
Skeleton-based action recognition has always been an important research topic of computer vision since the skeleton data is more robust to illumination and rotation. Traditional action recognition methods mainly rely on manual features. Among those methods, the skeleton feature representation modeled on Lie group can effectively describe the three-dimensional geometric relationship between joints. In recent years, deep learning methods such as Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN) and Long Short-Term Memory Networks (LSTM) have also achieved good performance in action recognition. In order to obtain more spatio-temporal information, we combine manual features and deep learning methods to propose a deep neural network LS-LieNet with CNN and Bi-direction LSTM (Bi-LSTM) based on the LieNet [1] network. First, the LS-LieNet network inputs the extracted Lie group representation of skeleton into a special CNN network which is designed for Lie group. Second, the transformed Lie algebra features are fed into the Bi-LSTM network before the fully connected layer of CNN. Then, the predicted labels and scores of the two network softmax layers are merged to effectively recognize the action. The experiment results on the standard 3D human action dataset show that the proposed LS-LieNet can efficiently improve the accuracy of action recognition.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.