Scalable detection of partial near-duplicate videos by visual-temporal consistency

Following the exponential growth of social media, there now exist huge repositories of videos online. Among the huge volumes of videos, there exist large numbers of near-duplicate videos. Most existing techniques either focus on the fast retrieval of full copies or near-duplicates, or consider local...

Full description

Saved in:
Bibliographic Details
Main Authors: TAN, Hung-Khoon, NGO, Chong-wah, HONG, Richang, CHUA, Tat-Seng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2009
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/6530
https://ink.library.smu.edu.sg/context/sis_research/article/7533/viewcontent/1631272.1631295.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-7533
record_format dspace
spelling sg-smu-ink.sis_research-75332022-01-10T03:48:04Z Scalable detection of partial near-duplicate videos by visual-temporal consistency TAN, Hung-Khoon NGO, Chong-wah HONG, Richang CHUA, Tat-Seng Following the exponential growth of social media, there now exist huge repositories of videos online. Among the huge volumes of videos, there exist large numbers of near-duplicate videos. Most existing techniques either focus on the fast retrieval of full copies or near-duplicates, or consider localization in a heuristic manner. This paper considers the scalable detection and localization of partial near-duplicate videos by jointly considering visual similarity and temporal consistency. Temporal constraints are embedded into a network structure as directed edges. Through the structure, partial alignment is novelly converted into a network flow problem where highly efficient solutions exist. To precisely decide the boundaries of the overlapping segments, pair-wise constraints generated from keypoint matching can be added to the network to iteratively refine the localization result. We demonstrate the effectiveness of partial alignment for three different tasks. The first task links partial segments in fulllength movies to videos crawled from YouTube. The second task performs fast web video search, while the third performs near-duplicate shot and copy detection. The experimental result demonstrates the effectiveness and efficiency of the proposed method compared to state-of-the-art techniques. 2009-10-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/6530 info:doi/10.1145/1631272.1631295 https://ink.library.smu.edu.sg/context/sis_research/article/7533/viewcontent/1631272.1631295.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Network flow Partial near-duplicate Temporal graph Databases and Information Systems Data Storage Systems Graphics and Human Computer Interfaces
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Network flow
Partial near-duplicate
Temporal graph
Databases and Information Systems
Data Storage Systems
Graphics and Human Computer Interfaces
spellingShingle Network flow
Partial near-duplicate
Temporal graph
Databases and Information Systems
Data Storage Systems
Graphics and Human Computer Interfaces
TAN, Hung-Khoon
NGO, Chong-wah
HONG, Richang
CHUA, Tat-Seng
Scalable detection of partial near-duplicate videos by visual-temporal consistency
description Following the exponential growth of social media, there now exist huge repositories of videos online. Among the huge volumes of videos, there exist large numbers of near-duplicate videos. Most existing techniques either focus on the fast retrieval of full copies or near-duplicates, or consider localization in a heuristic manner. This paper considers the scalable detection and localization of partial near-duplicate videos by jointly considering visual similarity and temporal consistency. Temporal constraints are embedded into a network structure as directed edges. Through the structure, partial alignment is novelly converted into a network flow problem where highly efficient solutions exist. To precisely decide the boundaries of the overlapping segments, pair-wise constraints generated from keypoint matching can be added to the network to iteratively refine the localization result. We demonstrate the effectiveness of partial alignment for three different tasks. The first task links partial segments in fulllength movies to videos crawled from YouTube. The second task performs fast web video search, while the third performs near-duplicate shot and copy detection. The experimental result demonstrates the effectiveness and efficiency of the proposed method compared to state-of-the-art techniques.
format text
author TAN, Hung-Khoon
NGO, Chong-wah
HONG, Richang
CHUA, Tat-Seng
author_facet TAN, Hung-Khoon
NGO, Chong-wah
HONG, Richang
CHUA, Tat-Seng
author_sort TAN, Hung-Khoon
title Scalable detection of partial near-duplicate videos by visual-temporal consistency
title_short Scalable detection of partial near-duplicate videos by visual-temporal consistency
title_full Scalable detection of partial near-duplicate videos by visual-temporal consistency
title_fullStr Scalable detection of partial near-duplicate videos by visual-temporal consistency
title_full_unstemmed Scalable detection of partial near-duplicate videos by visual-temporal consistency
title_sort scalable detection of partial near-duplicate videos by visual-temporal consistency
publisher Institutional Knowledge at Singapore Management University
publishDate 2009
url https://ink.library.smu.edu.sg/sis_research/6530
https://ink.library.smu.edu.sg/context/sis_research/article/7533/viewcontent/1631272.1631295.pdf
_version_ 1770575982469578752