Learning decoupled models for cross-modal generation
Cross-modal generation is playing an important role in translating information between different data modalities, such as image, video and text. Two representative tasks under the cross-modal generation umbrella are visual-to-text generation and text-to-visual generation. For the visual-to-text gene...
Saved in:
Main Author: | Wang, Hao |
---|---|
Other Authors: | Miao Chun Yan |
Format: | Thesis-Doctor of Philosophy |
Language: | English |
Published: |
Nanyang Technological University
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/169609 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
Similar Items
-
A decoupled learning framework for contrastive learning
by: Xu, Yicheng
Published: (2022) -
Cross-modal graph with meta concepts for video captioning
by: Wang, Hao, et al.
Published: (2022) -
Paired cross-modal data augmentation for fine-grained image-to-text retrieval
by: Wang, Hao, et al.
Published: (2023) -
Audio captioning and retrieval with improved cross-modal objectives
by: Koh, Andrew Jin Jie
Published: (2023) -
JALAD : joint accuracy- and latency-aware deep structure decoupling for edge-cloud execution
by: Li, Hongshan, et al.
Published: (2020)