Abstract

Temporal consistency stands as a vital property in semantic video retrieval. Few research studies can exploit this useful property. Most of the used methods in those studies depend on rules defined by experts and use ground-truth annotation. The Ground-truth annotation is time-consuming, labor intensive and domain specific. Additionally, it involves a limited number of annotated concepts and a limited number of annotated shots. Video concepts have interrelated relations, so the extracted temporal rules from ground-truth annotation are often inaccurate and incomplete. However, concept detection score data are a huge high-dimensional continuous-valued dataset and generated automatically. Temporal association rules algorithms are efficient methods in revealing the temporal relations, but they have some limitations when applied to high-dimensional and continuous-valued data. These constraints have led to a lack of research used temporal association rules. So, we propose a novel framework to encode the high-dimensional continuous-valued concept detection scores data into a single stream of numbers without loss of important information and to predict the neighbouring shots’ behavior by generating temporal association rules. Experiments on TRECVID 2010 dataset show that the proposed framework is both efficient and effective in encoding the dataset which reduces the dimensionality of the dataset matrix from 130×150000 dimensions to 130×1 dimensions without loss of important information and in predicting the behavior of neighbouring shots, the number of which can be 10 or more, using the extracted temporal rules.

Highlights

  • Tremendous growth in digital devices and digital media has led to the capture and storage of a huge amount of digital videos

  • This paper models and automates a framework to reduce the volume of video concept detection score data and to extract a compact representation of the temporal concept rules

  • The temporal rules are extracted from the stream of cluster numbers that resulted from the Gaussian mixture model clustering algorithm being applied to the data that were dimension reduced using Principle Component Analysis (PCA)

Read more

Summary

Introduction

Tremendous growth in digital devices and digital media has led to the capture and storage of a huge amount of digital videos. This paper models and automates a framework to reduce the volume of video concept detection score data and to extract a compact representation of the temporal concept rules. These rules predict the behavior of the neighboring shots based on the current and the previous shots’ behavior. Much research has been done on methods for discretizing or categorizing data to minimize the loss of information when converting data into the binary form, such methods increase the data dimensionality and do not prevent data loss To solve these difficulties, we apply Principle Component Analysis (PCA) in our method to compress the concept detection score matrix without loss of data.

Related Work
Experimental Setup
Conclusion and Future Work
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.