Abstract
Recent interest in e-learning and distance education services has significantly increased the amount of lecture video data in public and institutional repositories. In their current forms, users can browse in these collections using meta-data-based search queries such as course name, description, instructor and syllabus. However, lecture video entries have rich contents, including image, text and speech, which can not be easily represented by meta-data annotations. Therefore, there is an emerging need to develop tools that will automatically annotate lecture videos to facilitate more targeted search. A simple way to realize this is to classify lectures into known categories. With this objective, this paper presents a method for classifying videos based on extracted text content in several semantic levels. The method is based on Bidirectional Long-Short Term Memory (Bi-LSTM) applied on word embedding vectors of text content extracted by Optical Character Recognition (OCR). This approach can outperform conventional machine learning models and provide a useful solution for automatic lecture video annotation to support online education.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.