Abstract

In this paper we propose a way of using depth maps transformed into 3D point clouds to classify human activities. The activities are described as time sequences of feature vectors based on the Viewpoint Feature Histogram descriptor (VFH) computed using the Point Cloud Library. Recognition is performed by two types of classifiers: (i) k-NN nearest neighbors’ classifier with Dynamic Time Warping measure, (ii) bidirectional long short-term memory (BiLSTM) deep learning networks. Reduction of classification time for the k-NN by introducing a two tier model and improvement of BiLSTM-based classification via transfer learning and combining multiple networks by fuzzy integral are discussed. Our classification results obtained on two representative datasets: University of Texas at Dallas Multimodal Human Action Dataset and Mining Software Repositories Action 3D Dataset are comparable or better than the current state of the art.

Highlights

  • One of the most important tasks of human-computer interfaces is the interpretation of people’s behavior

  • In this paper we propose a way of using depth maps transformed into 3D point clouds to classify human activities

  • Complementing the mentioned applications related to hand gestures, these results can be seen as an argument for using the Viewpoint Feature Histogram descriptor (VFH) point cloud descriptors for people’s activity recognition

Read more

Summary

Introduction

One of the most important tasks of human-computer interfaces is the interpretation of people’s behavior. One of the best known devices for acquiring depth maps is the Microsoft KinectTM sensor. The increase in the popularity of the KinectTM and ToF cameras has greatly contributed to an increased interest in using depth maps. This data can be considered as an aid, and as the main source of information. In this paper we propose a way of using depth maps transformed into 3D point clouds to classify human activities. We describe the activities as time sequences of feature vectors based on the Viewpoint. Feature Histogram descriptor (VFH) computed using the point cloud library (PCL) [7].

Related Work and Contribution
Method
Values
Classification
BiLSTM
UTD Multimodal HUMAN Action Dataset
MSR-Action 3D Dataset
Activity Recognition System
Experiments
Activity Recognition Using DTW
Proposed Method
12. Recognition
Activity Recognition Using the BiLSTM Network
The use of BiLSTM Networks Fusion Using the Fuzzy Integral Method
Conclusions and Future Work
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.