(Un)likelihood training for interpretable embedding
Cross-modal representation learning has become a new normal for bridging the semantic gap between text and visual data. Learning modality agnostic representations in a continuous latent space, however, is often treated as a black-box data-driven training process. It is well known that the effectiven...
Saved in:
Main Authors: | WU, Jiaxin, NGO, Chong-wah, CHAN, Wing-Kwong, HOU, Zhijian |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2023
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/9819 https://ink.library.smu.edu.sg/context/sis_research/article/10819/viewcontent/2207.00282v3.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
Similar Items
-
Interpretable embedding for ad-hoc video search
by: WU, Jiaxin, et al.
Published: (2020) -
Improving interpretable embeddings for ad-hoc video search with generative captions and multi-word concept bank
by: WU, Jiaxin, et al.
Published: (2024) -
NEURAL NETWORK REPRESENTATION SIMILARITY REVISITED
by: WANG YUHUI
Published: (2024) -
SQL-like interpretable interactive video search
by: WU, Jiaxin, et al.
Published: (2021) -
Cross-modal recipe retrieval with stacked attention model
by: CHEN, Jing-Jing, et al.
Published: (2018)