Abstract

Dynamic texture analysis has been the focus of intensive research in recent years. Thus, this paper presents an innovative and highly discriminative dynamic texture analysis method, whose signature is composed of the weights of the output layer of a randomized neural network after a training procedure. This training is performed by using the pixels of slices of each orthogonal plane of the video (XY, YT, and XT) as input feature vectors and corresponding output labels. The obtained video signature provided an accuracy of 97.05%, 98.54%, 97.74% and 96.51% on the UCLA-50 classes, UCLA-9 classes, UCLA-8 classes and Dyntex++, respectively. These results, when compared to other dynamic texture analysis methods, demonstrate that our descriptors are very effective and that our proposed approach can contribute significantly to the field of dynamic texture analysis.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.