Deep-learning based affective video analysis and synthesis
The major challenge in computational creativity within the context of audio-visual analysis, is the difficulty in extracting high quality content from large quantities of video footage. Current development focuses on using submodular optimization of frame-based quality-aware relevance model to creat...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2020
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/144505 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-144505 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1445052020-11-10T04:54:56Z Deep-learning based affective video analysis and synthesis Tan, Christopher Say Wei Yu Han School of Computer Science and Engineering han.yu@ntu.edu.sg Engineering::Computer science and engineering The major challenge in computational creativity within the context of audio-visual analysis, is the difficulty in extracting high quality content from large quantities of video footage. Current development focuses on using submodular optimization of frame-based quality-aware relevance model to create summaries which are both diverse and representative of the entire video footage. Our work complements on existing work on query-adaptive video summarization, where we implement the Natural Language Toolkit and Rapid Automatic Keyword Extraction algorithm to extract keywords for query generation. The query is used in the Quality-Aware Relevance Estimation model for thumbnail selection. The generated thumbnails will identify key scenes in the video footage which will be subsequently summarized and merged by weighted sampling of the key scenes to the length of a short summary. We found that our video summary has more related scenes, higher average similarity score with key words compared to baseline, and it also improves on the average qualitative aspects of the summary. Bachelor of Engineering (Computer Science) 2020-11-10T04:54:56Z 2020-11-10T04:54:56Z 2020 Final Year Project (FYP) https://hdl.handle.net/10356/144505 en SCE19-0630 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering |
spellingShingle |
Engineering::Computer science and engineering Tan, Christopher Say Wei Deep-learning based affective video analysis and synthesis |
description |
The major challenge in computational creativity within the context of audio-visual analysis, is the difficulty in extracting high quality content from large quantities of video footage. Current development focuses on using submodular optimization of frame-based quality-aware relevance model to create summaries which are both diverse and representative of the entire video footage. Our work complements on existing work on query-adaptive video summarization, where we implement the Natural Language Toolkit and Rapid Automatic Keyword Extraction algorithm to extract keywords for query generation. The query is used in the Quality-Aware Relevance Estimation model for thumbnail selection. The generated thumbnails will identify key scenes in the video footage which will be subsequently summarized and merged by weighted sampling of the key scenes to the length of a short summary. We found that our video summary has more related scenes, higher average similarity score with key words compared to baseline, and it also improves on the average qualitative aspects of the summary. |
author2 |
Yu Han |
author_facet |
Yu Han Tan, Christopher Say Wei |
format |
Final Year Project |
author |
Tan, Christopher Say Wei |
author_sort |
Tan, Christopher Say Wei |
title |
Deep-learning based affective video analysis and synthesis |
title_short |
Deep-learning based affective video analysis and synthesis |
title_full |
Deep-learning based affective video analysis and synthesis |
title_fullStr |
Deep-learning based affective video analysis and synthesis |
title_full_unstemmed |
Deep-learning based affective video analysis and synthesis |
title_sort |
deep-learning based affective video analysis and synthesis |
publisher |
Nanyang Technological University |
publishDate |
2020 |
url |
https://hdl.handle.net/10356/144505 |
_version_ |
1686109392079945728 |