Improving interpretable embeddings for ad-hoc video search with generative captions and multi-word concept bank
Aligning a user query and video clips in cross-modal latent space and that with semantic concepts are two mainstream approaches for ad-hoc video search (AVS). However, the effectiveness of existing approaches is bottlenecked by the small sizes of available video-text datasets and the low quality of...
Saved in:
Main Authors: | WU, Jiaxin, NGO, Chong-wah, CHAN, Wing-Kwong |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2024
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/9288 https://ink.library.smu.edu.sg/context/sis_research/article/10288/viewcontent/2404.06173v1.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
Similar Items
-
Interpretable embedding for ad-hoc video search
by: WU, Jiaxin, et al.
Published: (2020) -
SQL-like interpretable interactive video search
by: WU, Jiaxin, et al.
Published: (2021) -
Building descriptive and discriminative visual codebook for large-scale image applications
by: Tian, Q., et al.
Published: (2016) -
Morphologically-aware vocabulary reduction of word embeddings
by: CHIA, Chong Cher, et al.
Published: (2023) -
Fusion of multimodal embeddings for ad-hoc video search
by: FRANCIS, Danny, et al.
Published: (2019)