Cross-Modal Self-Taught Hashing for large-scale image retrieval

Cross-modal hashing integrates the advantages of traditional cross-modal retrieval and hashing, it can solve large-scale cross-modal retrieval effectively and efficiently. However, existing cross-modal hashing methods rely on either labeled training data, or lack semantic analysis. In this paper, we...

Full description

Saved in:
Bibliographic Details
Main Authors: XIE, Liang, ZHU, Lei, PAN, Peng, LU, Yansheng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2016
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/3587
https://ink.library.smu.edu.sg/context/sis_research/article/4588/viewcontent/cross_modal__1_.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-4588
record_format dspace
spelling sg-smu-ink.sis_research-45882020-01-12T11:58:21Z Cross-Modal Self-Taught Hashing for large-scale image retrieval XIE, Liang ZHU, Lei PAN, Peng LU, Yansheng Cross-modal hashing integrates the advantages of traditional cross-modal retrieval and hashing, it can solve large-scale cross-modal retrieval effectively and efficiently. However, existing cross-modal hashing methods rely on either labeled training data, or lack semantic analysis. In this paper, we propose Cross-Modal Self-Taught Hashing (CMSTH) for large-scale cross-modal and unimodal image retrieval. CMSTH can effectively capture the semantic correlation from unlabeled training data. Its learning process contains three steps: first we propose Hierarchical Multi-Modal Topic Learning (HMMTL) to detect multi-modal topics with semantic information. Then we use Robust Matrix Factorization (RMF) to transfer the multi-modal topics to hash codes which are more suited to quantization, and these codes form a unified hash space. Finally we learn hash functions to project all modalities into the unified hash space. Experimental results on two web image datasets demonstrate the effectiveness of CMSTH compared to representative cross-modal and unimodal hashing methods. 2016-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/3587 info:doi/10.1016/j.sigpro.2015.10.010 https://ink.library.smu.edu.sg/context/sis_research/article/4588/viewcontent/cross_modal__1_.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Cross-modal hashing Image retreival Self-taught learning Semantic correlation Graphics and Human Computer Interfaces Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Cross-modal hashing
Image retreival
Self-taught learning
Semantic correlation
Graphics and Human Computer Interfaces
Software Engineering
spellingShingle Cross-modal hashing
Image retreival
Self-taught learning
Semantic correlation
Graphics and Human Computer Interfaces
Software Engineering
XIE, Liang
ZHU, Lei
PAN, Peng
LU, Yansheng
Cross-Modal Self-Taught Hashing for large-scale image retrieval
description Cross-modal hashing integrates the advantages of traditional cross-modal retrieval and hashing, it can solve large-scale cross-modal retrieval effectively and efficiently. However, existing cross-modal hashing methods rely on either labeled training data, or lack semantic analysis. In this paper, we propose Cross-Modal Self-Taught Hashing (CMSTH) for large-scale cross-modal and unimodal image retrieval. CMSTH can effectively capture the semantic correlation from unlabeled training data. Its learning process contains three steps: first we propose Hierarchical Multi-Modal Topic Learning (HMMTL) to detect multi-modal topics with semantic information. Then we use Robust Matrix Factorization (RMF) to transfer the multi-modal topics to hash codes which are more suited to quantization, and these codes form a unified hash space. Finally we learn hash functions to project all modalities into the unified hash space. Experimental results on two web image datasets demonstrate the effectiveness of CMSTH compared to representative cross-modal and unimodal hashing methods.
format text
author XIE, Liang
ZHU, Lei
PAN, Peng
LU, Yansheng
author_facet XIE, Liang
ZHU, Lei
PAN, Peng
LU, Yansheng
author_sort XIE, Liang
title Cross-Modal Self-Taught Hashing for large-scale image retrieval
title_short Cross-Modal Self-Taught Hashing for large-scale image retrieval
title_full Cross-Modal Self-Taught Hashing for large-scale image retrieval
title_fullStr Cross-Modal Self-Taught Hashing for large-scale image retrieval
title_full_unstemmed Cross-Modal Self-Taught Hashing for large-scale image retrieval
title_sort cross-modal self-taught hashing for large-scale image retrieval
publisher Institutional Knowledge at Singapore Management University
publishDate 2016
url https://ink.library.smu.edu.sg/sis_research/3587
https://ink.library.smu.edu.sg/context/sis_research/article/4588/viewcontent/cross_modal__1_.pdf
_version_ 1770573337131483136