Action search by example using randomized visual vocabularies

Because actions can be small video objects, it is a challenging problem to search for similar actions in crowded and dynamic scenes when a single query example is provided. We propose a fast action search method that can efficiently locate similar actions spatiotemporally. Both the query action and...

Full description

Saved in:
Bibliographic Details
Main Authors: Yuan, Junsong, Liu, Zicheng, Yu, Gang
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2013
Subjects:
Online Access:https://hdl.handle.net/10356/100686
http://hdl.handle.net/10220/17968
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-100686
record_format dspace
spelling sg-ntu-dr.10356-1006862020-03-07T14:00:31Z Action search by example using randomized visual vocabularies Yuan, Junsong Liu, Zicheng Yu, Gang School of Electrical and Electronic Engineering DRNTU::Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision Because actions can be small video objects, it is a challenging problem to search for similar actions in crowded and dynamic scenes when a single query example is provided. We propose a fast action search method that can efficiently locate similar actions spatiotemporally. Both the query action and the video datasets are characterized by spatio-temporal interest points. Instead of using a unified visual vocabulary to index all interest points in the database, we propose randomized visual vocabularies to enable fast and robust interest point matching. To accelerate action localization, we have developed a coarse-to-fine video subvolume search scheme, which is several orders of magnitude faster than the existing spatio-temporal branch and bound search. Our experiments on cross-dataset action search show promising results when compared with the state of the arts. Additional experiments on a 5-h versatile video dataset validate the efficiency of our method, where an action search can be finished in just 37.6 s on a regular desktop machine. Accepted version 2013-12-02T07:24:59Z 2019-12-06T20:26:38Z 2013-12-02T07:24:59Z 2019-12-06T20:26:38Z 2013 2013 Journal Article Yu, G., Yuan, J., & Liu, Z. (2013). Action Search by Example using Randomized Visual Vocabularies. IEEE Transactions on Image Processing, 22(1), 377-390. 1057-7149 https://hdl.handle.net/10356/100686 http://hdl.handle.net/10220/17968 10.1109/TIP.2012.2216273 en IEEE transactions on image processing © 2013 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. The published version is available at: [http://dx.doi.org/10.1109/TIP.2012.2216273]. 15 p. application/pdf
institution Nanyang Technological University
building NTU Library
country Singapore
collection DR-NTU
language English
topic DRNTU::Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
spellingShingle DRNTU::Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
Yuan, Junsong
Liu, Zicheng
Yu, Gang
Action search by example using randomized visual vocabularies
description Because actions can be small video objects, it is a challenging problem to search for similar actions in crowded and dynamic scenes when a single query example is provided. We propose a fast action search method that can efficiently locate similar actions spatiotemporally. Both the query action and the video datasets are characterized by spatio-temporal interest points. Instead of using a unified visual vocabulary to index all interest points in the database, we propose randomized visual vocabularies to enable fast and robust interest point matching. To accelerate action localization, we have developed a coarse-to-fine video subvolume search scheme, which is several orders of magnitude faster than the existing spatio-temporal branch and bound search. Our experiments on cross-dataset action search show promising results when compared with the state of the arts. Additional experiments on a 5-h versatile video dataset validate the efficiency of our method, where an action search can be finished in just 37.6 s on a regular desktop machine.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Yuan, Junsong
Liu, Zicheng
Yu, Gang
format Article
author Yuan, Junsong
Liu, Zicheng
Yu, Gang
author_sort Yuan, Junsong
title Action search by example using randomized visual vocabularies
title_short Action search by example using randomized visual vocabularies
title_full Action search by example using randomized visual vocabularies
title_fullStr Action search by example using randomized visual vocabularies
title_full_unstemmed Action search by example using randomized visual vocabularies
title_sort action search by example using randomized visual vocabularies
publishDate 2013
url https://hdl.handle.net/10356/100686
http://hdl.handle.net/10220/17968
_version_ 1681045093803360256