Mobile media tagging and sharing

Huge advancement in mobile and network technology has open various opportunities for development in mobile media application. One such area involves content-based visual information retrieval (CVIR), where mobile users are able to make a search using images rather than words. The long-term goal is t...

Full description

Saved in:
Bibliographic Details
Main Author: Lam, Jasmine Xin Yi
Other Authors: Yap Kim Hui
Format: Final Year Project
Language:English
Published: 2013
Subjects:
Online Access:http://hdl.handle.net/10356/52600
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-52600
record_format dspace
spelling sg-ntu-dr.10356-526002023-07-07T17:29:11Z Mobile media tagging and sharing Lam, Jasmine Xin Yi Yap Kim Hui School of Electrical and Electronic Engineering DRNTU::Engineering Huge advancement in mobile and network technology has open various opportunities for development in mobile media application. One such area involves content-based visual information retrieval (CVIR), where mobile users are able to make a search using images rather than words. The long-term goal is to integrate different methods of tagging and searching of images to improve performance. However, this final year project aims to build the fundamental foundation for the long-term project first by experimenting on the existing image recognition tools and explore different techniques to achieve high performance accuracy with the shortest cost (computational time). To narrow the scope, the area of focus is on Landmark recognition. The Singapore Landmark Database is utilized. It consists of 40 categories and 13409 images and has been reorganised with real-time application in mind. While both Matlab and OpenCV codes are used, on the other hand OpenCV is mainly utilized to allow easier system integration for future works. A Bag-of-Words framework is adopted for the image recognition tools. For feature extraction the Scale-based Invariant Feature Transform (SIFT) is used, while both the hierarchical k-means and scalable vocabulary tree are used for clustering and machine learning respectively. Two different sampling methods are experimented on; the Dense SIFT (dense- sampling) and the Key-Point SIFT (key-point sampling). The experimental results conclude that Key-Point SIFT is better performing with an increase in recognition rate (>19%) and has a faster computational time (>20%) for landmark recognition. Two additional experiments were conducted; the first on Geometric Verification (GV) and second on Saliency Mapping integration. The results conclude that both GV and Saliency Mapping allow better performance (1~2%), but at the expense of computational time (increase in GV >2s; increase in Saliency <0.03s). Future works may include the expansion of the Singapore Landmark database and integration of GPS into the stored image content to increase performance accuracy. Bachelor of Engineering 2013-05-21T02:15:29Z 2013-05-21T02:15:29Z 2013 2013 Final Year Project (FYP) http://hdl.handle.net/10356/52600 en Nanyang Technological University 52 p. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic DRNTU::Engineering
spellingShingle DRNTU::Engineering
Lam, Jasmine Xin Yi
Mobile media tagging and sharing
description Huge advancement in mobile and network technology has open various opportunities for development in mobile media application. One such area involves content-based visual information retrieval (CVIR), where mobile users are able to make a search using images rather than words. The long-term goal is to integrate different methods of tagging and searching of images to improve performance. However, this final year project aims to build the fundamental foundation for the long-term project first by experimenting on the existing image recognition tools and explore different techniques to achieve high performance accuracy with the shortest cost (computational time). To narrow the scope, the area of focus is on Landmark recognition. The Singapore Landmark Database is utilized. It consists of 40 categories and 13409 images and has been reorganised with real-time application in mind. While both Matlab and OpenCV codes are used, on the other hand OpenCV is mainly utilized to allow easier system integration for future works. A Bag-of-Words framework is adopted for the image recognition tools. For feature extraction the Scale-based Invariant Feature Transform (SIFT) is used, while both the hierarchical k-means and scalable vocabulary tree are used for clustering and machine learning respectively. Two different sampling methods are experimented on; the Dense SIFT (dense- sampling) and the Key-Point SIFT (key-point sampling). The experimental results conclude that Key-Point SIFT is better performing with an increase in recognition rate (>19%) and has a faster computational time (>20%) for landmark recognition. Two additional experiments were conducted; the first on Geometric Verification (GV) and second on Saliency Mapping integration. The results conclude that both GV and Saliency Mapping allow better performance (1~2%), but at the expense of computational time (increase in GV >2s; increase in Saliency <0.03s). Future works may include the expansion of the Singapore Landmark database and integration of GPS into the stored image content to increase performance accuracy.
author2 Yap Kim Hui
author_facet Yap Kim Hui
Lam, Jasmine Xin Yi
format Final Year Project
author Lam, Jasmine Xin Yi
author_sort Lam, Jasmine Xin Yi
title Mobile media tagging and sharing
title_short Mobile media tagging and sharing
title_full Mobile media tagging and sharing
title_fullStr Mobile media tagging and sharing
title_full_unstemmed Mobile media tagging and sharing
title_sort mobile media tagging and sharing
publishDate 2013
url http://hdl.handle.net/10356/52600
_version_ 1772825912840028160