Detection of visual attention regions in images and videos

The explosive growth of multimedia content and advances in the development of hardware with multimedia functionalities call for techniques to enable users to access such content anywhere and anytime and with similarly pleasing experience each time. This requires intelligent search, transmission, ana...

Full description

Saved in:
Bibliographic Details
Main Author: Hu, Yiqun
Other Authors: Chia Liang Tien
Format: Theses and Dissertations
Language:English
Published: 2009
Subjects:
Online Access:https://hdl.handle.net/10356/18864
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-18864
record_format dspace
spelling sg-ntu-dr.10356-188642023-03-04T00:40:12Z Detection of visual attention regions in images and videos Hu, Yiqun Chia Liang Tien Deepu Rajan School of Computer Engineering DRNTU::Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision The explosive growth of multimedia content and advances in the development of hardware with multimedia functionalities call for techniques to enable users to access such content anywhere and anytime and with similarly pleasing experience each time. This requires intelligent search, transmission, analysis and display of multimedia data. However, in addition to the data being very large in size, it is inherently complex due to the variety of features (color, texture, shapes, motion, etc.) that it contains. The challenge then is to detect information front the clutter for further processing. The relevant information is the visual attention region (VAR) whose detection in images and videos is the topic of this dissertation. The bottom-up model for detecting VAR in an image involves generation of a saliency map that highlights contrasts in features like color, intensity and orientation. The saliency map itself is obtained through a combination of each feature map that highlights the contrast for that particular feature. We investigate the process of good selection and proper combination strategies for the features. We propose a novel Composite Saliency Indicator (CSI) to determine the contribution of each feature map to the salient region. CS1 is designed to capture the spatial compactness as well as the density of candidate regions in the feature maps. We also propose a Context Suppression Model that provides a measure to determine similarity among candidate attention regions in a feature map. This measure is used to find a suppression factor for a particular patch in the scene, which is then used to highlight actual attention regions. We also demonstrate an application in multimedia adaptation that benefits from the improved VAR detection. DOCTOR OF PHILOSOPHY (SCE) 2009-07-20T07:37:46Z 2009-07-20T07:37:46Z 2008 2008 Thesis Hu, Y. Q. (2008). Detection of visual attention regions in images and videos. Doctoral thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/18864 10.32657/10356/18864 en 163 p. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic DRNTU::Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
spellingShingle DRNTU::Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
Hu, Yiqun
Detection of visual attention regions in images and videos
description The explosive growth of multimedia content and advances in the development of hardware with multimedia functionalities call for techniques to enable users to access such content anywhere and anytime and with similarly pleasing experience each time. This requires intelligent search, transmission, analysis and display of multimedia data. However, in addition to the data being very large in size, it is inherently complex due to the variety of features (color, texture, shapes, motion, etc.) that it contains. The challenge then is to detect information front the clutter for further processing. The relevant information is the visual attention region (VAR) whose detection in images and videos is the topic of this dissertation. The bottom-up model for detecting VAR in an image involves generation of a saliency map that highlights contrasts in features like color, intensity and orientation. The saliency map itself is obtained through a combination of each feature map that highlights the contrast for that particular feature. We investigate the process of good selection and proper combination strategies for the features. We propose a novel Composite Saliency Indicator (CSI) to determine the contribution of each feature map to the salient region. CS1 is designed to capture the spatial compactness as well as the density of candidate regions in the feature maps. We also propose a Context Suppression Model that provides a measure to determine similarity among candidate attention regions in a feature map. This measure is used to find a suppression factor for a particular patch in the scene, which is then used to highlight actual attention regions. We also demonstrate an application in multimedia adaptation that benefits from the improved VAR detection.
author2 Chia Liang Tien
author_facet Chia Liang Tien
Hu, Yiqun
format Theses and Dissertations
author Hu, Yiqun
author_sort Hu, Yiqun
title Detection of visual attention regions in images and videos
title_short Detection of visual attention regions in images and videos
title_full Detection of visual attention regions in images and videos
title_fullStr Detection of visual attention regions in images and videos
title_full_unstemmed Detection of visual attention regions in images and videos
title_sort detection of visual attention regions in images and videos
publishDate 2009
url https://hdl.handle.net/10356/18864
_version_ 1759854421790949376