Paired cross-modal data augmentation for fine-grained image-to-text retrieval

This paper investigates an open research problem of generating text-image pairs to improve the training of fine-grained image-to-text cross-modal retrieval task, and proposes a novel framework for paired data augmentation by uncovering the hidden semantic information of StyleGAN2 model. Specific...

Full description

Saved in:
Bibliographic Details
Main Authors: Wang, Hao, Lin, Guosheng, Hoi, Steven C. H., Miao, Chunyan
Other Authors: School of Computer Science and Engineering
Format: Conference or Workshop Item
Language:English
Published: 2023
Subjects:
Online Access:https://hdl.handle.net/10356/164145
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-164145
record_format dspace
spelling sg-ntu-dr.10356-1641452023-06-02T15:35:42Z Paired cross-modal data augmentation for fine-grained image-to-text retrieval Wang, Hao Lin, Guosheng Hoi, Steven C. H. Miao, Chunyan School of Computer Science and Engineering 30th ACM International Conference on Multimedia (MM 2022) Engineering::Computer science and engineering Image-to-Text Retrieval Computing Methodologies This paper investigates an open research problem of generating text-image pairs to improve the training of fine-grained image-to-text cross-modal retrieval task, and proposes a novel framework for paired data augmentation by uncovering the hidden semantic information of StyleGAN2 model. Specifically, we first train a StyleGAN2 model on the given dataset. We then project the real images back to the latent space of StyleGAN2 to obtain the latent codes. To make the generated images manipulatable, we further introduce a latent space alignment module to learn the alignment between StyleGAN2 latent codes and the corresponding textual caption features. When we do online paired data augmentation, we first generate augmented text through random token replacement, then pass the augmented text into the latent space alignment module to output the latent codes, which are finally fed to StyleGAN2 to generate the augmented images. We evaluate the efficacy of our augmented data approach on two public cross-modal retrieval datasets, in which the promising experimental results demonstrate the augmented text-image pair data can be trained together with the original data to boost the image-to-text cross-modal retrieval performance. AI Singapore Ministry of Education (MOE) Ministry of Health (MOH) National Research Foundation (NRF) Submitted/Accepted version This research is supported, in part, by the National Research Foun- dation (NRF), Singapore under its AI Singapore Programme (AISG Award No: AISG-GC-2019-003) and under its NRF Investigator- ship Programme (NRFI Award No. NRF-NRFI05-2019-0002). This research is supported, in part, by the Singapore Ministry of Health under its National Innovation Challenge on Active and Confident Ageing (NIC Project No. MOH/NIC/HAIG03/2017). This research is also supported by the National Research Foundation, Singapore under its AI Singapore Programme (AISG Award No: AISG-RP-2018-003), and the MOE AcRF Tier-1 research grant: RG95/20. 2023-01-06T06:05:12Z 2023-01-06T06:05:12Z 2022 Conference Paper Wang, H., Lin, G., Hoi, S. C. H. & Miao, C. (2022). Paired cross-modal data augmentation for fine-grained image-to-text retrieval. 30th ACM International Conference on Multimedia (MM 2022), 5517-5526. https://dx.doi.org/10.1145/3503161.3547809 9781450392037 https://hdl.handle.net/10356/164145 10.1145/3503161.3547809 5517 5526 en AISG-GC-2019-003 NRF-NRFI05-2019-0002 MOH/NIC/HAIG03/2017 AISG-RP-2018-003 RG95/20 © 2022 The owner/author(s). Publication rights licensed to ACM. All rights reserved. This paper was published in the Proceedings of 30th ACM International Conference on Multimedia (MM 2022) and is made available with permission of The owner/author(s). application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering
Image-to-Text Retrieval
Computing Methodologies
spellingShingle Engineering::Computer science and engineering
Image-to-Text Retrieval
Computing Methodologies
Wang, Hao
Lin, Guosheng
Hoi, Steven C. H.
Miao, Chunyan
Paired cross-modal data augmentation for fine-grained image-to-text retrieval
description This paper investigates an open research problem of generating text-image pairs to improve the training of fine-grained image-to-text cross-modal retrieval task, and proposes a novel framework for paired data augmentation by uncovering the hidden semantic information of StyleGAN2 model. Specifically, we first train a StyleGAN2 model on the given dataset. We then project the real images back to the latent space of StyleGAN2 to obtain the latent codes. To make the generated images manipulatable, we further introduce a latent space alignment module to learn the alignment between StyleGAN2 latent codes and the corresponding textual caption features. When we do online paired data augmentation, we first generate augmented text through random token replacement, then pass the augmented text into the latent space alignment module to output the latent codes, which are finally fed to StyleGAN2 to generate the augmented images. We evaluate the efficacy of our augmented data approach on two public cross-modal retrieval datasets, in which the promising experimental results demonstrate the augmented text-image pair data can be trained together with the original data to boost the image-to-text cross-modal retrieval performance.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Wang, Hao
Lin, Guosheng
Hoi, Steven C. H.
Miao, Chunyan
format Conference or Workshop Item
author Wang, Hao
Lin, Guosheng
Hoi, Steven C. H.
Miao, Chunyan
author_sort Wang, Hao
title Paired cross-modal data augmentation for fine-grained image-to-text retrieval
title_short Paired cross-modal data augmentation for fine-grained image-to-text retrieval
title_full Paired cross-modal data augmentation for fine-grained image-to-text retrieval
title_fullStr Paired cross-modal data augmentation for fine-grained image-to-text retrieval
title_full_unstemmed Paired cross-modal data augmentation for fine-grained image-to-text retrieval
title_sort paired cross-modal data augmentation for fine-grained image-to-text retrieval
publishDate 2023
url https://hdl.handle.net/10356/164145
_version_ 1772827845833261056