Abstract
In recent years, the recognition of activity is a daring task which helps elderly people, disabled patients and so on. The aim of this paper is to design a system for recognizing the human activity in egocentric video. In this research work, the various textural features like gray level co-occurrence matrix and local binary pattern and point feature speeded up robust features are retrieved from activity videos which is a proposed work and classifiers like probabilistic neural network, support vector machine (SVM), k nearest neighbor (kNN) and proposed SVM+kNN classifiers are used to classify the activity. Here, multimodal egocentric activity dataset is chosen as input. The performance results showed that the SVM+kNN classifier outperformed other classifiers.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.