Abstract
This Video summarization is a vital task in multimedia analysis, especially given the vast volume of video data in the digital age. Although deep learning methods have been extensively studied for this purpose, they often face challenges in efficiently processing long-duration videos. This paper tackles the issue of unsupervised video summarization by introducing a novel approach that selects a sparse subset of video frames to optimally represent the original video. The core concept involves training a deep summarizer network within a generative adversarial framework, which includes an autoencoder LSTM network as the summarizer and another LSTM network as the discriminator. The summarizer LSTM is designed to select key video frames and reconstruct the input video from these selected frames. Concurrently, the discriminator LSTM's role is to differentiate between the original video and its reconstruction. Through adversarial training between the summarizer and discriminator, combined with sparsity regularization, the network learns to produce optimal video summaries without requiring labeled data. Evaluations on several benchmark datasets indicate that this method delivers competitive performance compared to fully supervised state-of-the-art techniques, highlighting its effectiveness in unsupervised video summarization. Key Words: Event summarization, Critical information in videos, Surveillance systems ,Video analysis, Multimedia analysis, Deep learning, Unsupervised learning, Autoencoder LSTM, Long short-term memory network (LSTM)
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: INTERANTIONAL JOURNAL OF SCIENTIFIC RESEARCH IN ENGINEERING AND MANAGEMENT
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.