Summarizing rushes videos by motion, object, and event understanding
Rushes footages are considered as cheap gold mine with the potential for reuse in broadcasting and filmmaking industries. However, mining "gold" from unedited videos such as rushes is challenging as the reusable segments are buried in a large set of redundant information. In this paper, we...
Saved in:
Main Authors: | , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2012
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/6346 https://ink.library.smu.edu.sg/context/sis_research/article/7349/viewcontent/tmm12_wf.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-7349 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-73492021-11-23T04:05:03Z Summarizing rushes videos by motion, object, and event understanding WANG, Feng NGO, Chong-wah Rushes footages are considered as cheap gold mine with the potential for reuse in broadcasting and filmmaking industries. However, mining "gold" from unedited videos such as rushes is challenging as the reusable segments are buried in a large set of redundant information. In this paper, we propose a unified framework for stock footage classification and summarization to support video editors in navigating and organizing rushes videos. Our approach is composed of two steps. First, we employ motion features to filter the undesired camera motion and locate the stock footage. A hierarchical hidden Markov model (HHMM) is proposed to model the motion feature distribution and classify video segments into different categories to decide their potential for reuse. Second, we generate a short video summary to facilitate quick browsing of the stock footages by including the objects and events that are important for storytelling. For objects, we detect the presence of persons and moving objects. For events, we extract a set of features to detect and describe visual (motion activities and scene changes) and audio events (speech clips). A representability measure is then proposed to select the most representative video clips for video summarization. Our experiments show that the proposed HHMM significantly outperforms other methods based on SVM, FSM, and HMM. The automatically generated rushes summaries are also demonstrated to be easy-to-understand, containing little redundancy, and capable of including ground-truth objects and events with shorter durations and relatively pleasant rhythm based on the TRECVID 2007, 2008, and our subjective evaluations. 2012-02-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/6346 info:doi/10.1109/TMM.2011.2165531 https://ink.library.smu.edu.sg/context/sis_research/article/7349/viewcontent/tmm12_wf.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Motion analysis object and event understanding rushes video structuring video summarization Computer Sciences Graphics and Human Computer Interfaces |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Motion analysis object and event understanding rushes video structuring video summarization Computer Sciences Graphics and Human Computer Interfaces |
spellingShingle |
Motion analysis object and event understanding rushes video structuring video summarization Computer Sciences Graphics and Human Computer Interfaces WANG, Feng NGO, Chong-wah Summarizing rushes videos by motion, object, and event understanding |
description |
Rushes footages are considered as cheap gold mine with the potential for reuse in broadcasting and filmmaking industries. However, mining "gold" from unedited videos such as rushes is challenging as the reusable segments are buried in a large set of redundant information. In this paper, we propose a unified framework for stock footage classification and summarization to support video editors in navigating and organizing rushes videos. Our approach is composed of two steps. First, we employ motion features to filter the undesired camera motion and locate the stock footage. A hierarchical hidden Markov model (HHMM) is proposed to model the motion feature distribution and classify video segments into different categories to decide their potential for reuse. Second, we generate a short video summary to facilitate quick browsing of the stock footages by including the objects and events that are important for storytelling. For objects, we detect the presence of persons and moving objects. For events, we extract a set of features to detect and describe visual (motion activities and scene changes) and audio events (speech clips). A representability measure is then proposed to select the most representative video clips for video summarization. Our experiments show that the proposed HHMM significantly outperforms other methods based on SVM, FSM, and HMM. The automatically generated rushes summaries are also demonstrated to be easy-to-understand, containing little redundancy, and capable of including ground-truth objects and events with shorter durations and relatively pleasant rhythm based on the TRECVID 2007, 2008, and our subjective evaluations. |
format |
text |
author |
WANG, Feng NGO, Chong-wah |
author_facet |
WANG, Feng NGO, Chong-wah |
author_sort |
WANG, Feng |
title |
Summarizing rushes videos by motion, object, and event understanding |
title_short |
Summarizing rushes videos by motion, object, and event understanding |
title_full |
Summarizing rushes videos by motion, object, and event understanding |
title_fullStr |
Summarizing rushes videos by motion, object, and event understanding |
title_full_unstemmed |
Summarizing rushes videos by motion, object, and event understanding |
title_sort |
summarizing rushes videos by motion, object, and event understanding |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2012 |
url |
https://ink.library.smu.edu.sg/sis_research/6346 https://ink.library.smu.edu.sg/context/sis_research/article/7349/viewcontent/tmm12_wf.pdf |
_version_ |
1770575938983034880 |