Deep-learning based affective video analysis and synthesis

The major challenge in computational creativity within the context of audio-visual analysis, is the difficulty in extracting high quality content from large quantities of video footage. Current development focuses on using submodular optimization of frame-based quality-aware relevance model to creat...

وصف كامل

محفوظ في:
التفاصيل البيبلوغرافية
المؤلف الرئيسي: Tan, Christopher Say Wei
مؤلفون آخرون: Yu Han
التنسيق: Final Year Project
اللغة:English
منشور في: Nanyang Technological University 2020
الموضوعات:
الوصول للمادة أونلاين:https://hdl.handle.net/10356/144505
الوسوم: إضافة وسم
لا توجد وسوم, كن أول من يضع وسما على هذه التسجيلة!
المؤسسة: Nanyang Technological University
اللغة: English
الوصف
الملخص:The major challenge in computational creativity within the context of audio-visual analysis, is the difficulty in extracting high quality content from large quantities of video footage. Current development focuses on using submodular optimization of frame-based quality-aware relevance model to create summaries which are both diverse and representative of the entire video footage. Our work complements on existing work on query-adaptive video summarization, where we implement the Natural Language Toolkit and Rapid Automatic Keyword Extraction algorithm to extract keywords for query generation. The query is used in the Quality-Aware Relevance Estimation model for thumbnail selection. The generated thumbnails will identify key scenes in the video footage which will be subsequently summarized and merged by weighted sampling of the key scenes to the length of a short summary. We found that our video summary has more related scenes, higher average similarity score with key words compared to baseline, and it also improves on the average qualitative aspects of the summary.