Abstract

Video description (VD) aims to automatically generate descriptive natural language for videos. With its successful implementations and a broad range of applications, lots of work based on deep neural network (DNN) models have been put forward by researchers. This paper takes inspiration from an image caption model and develops an end-to-end VD model using long short-term memory (LSTM). Single video feature is fed to the first unit of LSTM decoder, and subsequent words of sentence are generated on previous predicted words. Experimental results on two publicly available datasets demonstrate that the performance of the proposed model outperforms that of baseline.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call