2D and 3D video scene text classification

Text detection and recognition is a challenging problem in document analysis due 10 the presence of the unpredictable nature of video texts, such as the variations of orientation, font and size, illumination effects, and even different 20/30 text shadows. In this paper, we propose a novel horizontal...

Full description

Saved in:
Bibliographic Details
Main Authors: Xu, J., Shivakumara, P., Lu, T., Tan, C.L.
Format: Conference or Workshop Item
Language:English
Published: 2014
Subjects:
Online Access:http://eprints.um.edu.my/13101/1/2d_and_3d_video_scene_text.pdf
http://eprints.um.edu.my/13101/
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Universiti Malaya
Language: English
id my.um.eprints.13101
record_format eprints
spelling my.um.eprints.131012015-03-24T06:14:06Z http://eprints.um.edu.my/13101/ 2D and 3D video scene text classification Xu, J. Shivakumara, P. Lu, T. Tan, C.L. T Technology (General) Text detection and recognition is a challenging problem in document analysis due 10 the presence of the unpredictable nature of video texts, such as the variations of orientation, font and size, illumination effects, and even different 20/30 text shadows. In this paper, we propose a novel horizontal and vertical symmetry feature by calculating the gradient direction and the gradient magnitude of each text candidate, which results in Potential Text Candidates (PTCs) after applying the k-means clustering algorithm on the gradient image of each input frame to verify PTC , we explore temporal information of video by proposing an iterative process that continuously verifies the PTCs of the first frame and the successive frames, until the process meets the converging criterion. This outputs Stable Potential Text Candidates (SPTCs). For each , PTC, the method obtains text representatives with the help of the edge image of the input frame. Then for each text representative, we divide it into four quadrants and check a new Mutual Nearest Neighbor Symmetry (MNNS) based on the dominant stroke width distances of the four quadrants. A voting method is finally proposed to clasify each text block as either 2D or 3D by counting the text representatives that satisfy MNNS. Experimental results on clasifying 2D and 3D text images are promising, and the result re further validated by text detection and recognition before clasification and after clasification with the exiting methods, respectively. 2014-08 Conference or Workshop Item PeerReviewed application/pdf en http://eprints.um.edu.my/13101/1/2d_and_3d_video_scene_text.pdf Xu, J. and Shivakumara, P. and Lu, T. and Tan, C.L. (2014) 2D and 3D video scene text classification. In: International Conference on Pattern Recognition (ICPR) , 24-28 Aug 2014, Stockholm, Sweden. (Submitted)
institution Universiti Malaya
building UM Library
collection Institutional Repository
continent Asia
country Malaysia
content_provider Universiti Malaya
content_source UM Research Repository
url_provider http://eprints.um.edu.my/
language English
topic T Technology (General)
spellingShingle T Technology (General)
Xu, J.
Shivakumara, P.
Lu, T.
Tan, C.L.
2D and 3D video scene text classification
description Text detection and recognition is a challenging problem in document analysis due 10 the presence of the unpredictable nature of video texts, such as the variations of orientation, font and size, illumination effects, and even different 20/30 text shadows. In this paper, we propose a novel horizontal and vertical symmetry feature by calculating the gradient direction and the gradient magnitude of each text candidate, which results in Potential Text Candidates (PTCs) after applying the k-means clustering algorithm on the gradient image of each input frame to verify PTC , we explore temporal information of video by proposing an iterative process that continuously verifies the PTCs of the first frame and the successive frames, until the process meets the converging criterion. This outputs Stable Potential Text Candidates (SPTCs). For each , PTC, the method obtains text representatives with the help of the edge image of the input frame. Then for each text representative, we divide it into four quadrants and check a new Mutual Nearest Neighbor Symmetry (MNNS) based on the dominant stroke width distances of the four quadrants. A voting method is finally proposed to clasify each text block as either 2D or 3D by counting the text representatives that satisfy MNNS. Experimental results on clasifying 2D and 3D text images are promising, and the result re further validated by text detection and recognition before clasification and after clasification with the exiting methods, respectively.
format Conference or Workshop Item
author Xu, J.
Shivakumara, P.
Lu, T.
Tan, C.L.
author_facet Xu, J.
Shivakumara, P.
Lu, T.
Tan, C.L.
author_sort Xu, J.
title 2D and 3D video scene text classification
title_short 2D and 3D video scene text classification
title_full 2D and 3D video scene text classification
title_fullStr 2D and 3D video scene text classification
title_full_unstemmed 2D and 3D video scene text classification
title_sort 2d and 3d video scene text classification
publishDate 2014
url http://eprints.um.edu.my/13101/1/2d_and_3d_video_scene_text.pdf
http://eprints.um.edu.my/13101/
_version_ 1643689463091560448