Question-guided hybrid convolution for visual question answering
In this paper, we propose a novel Question-Guided Hybrid Convolution (QGHC) network for Visual Question Answering (VQA). Most state-of-the-art VQA methods fuse the high-level textual and visual features from the neural network and abandon the visual spatial information when learning multi-modal feat...
Saved in:
Main Authors: | GAO, Peng, LU, Pan, LI, Hongsheng, LI, Shuang, LI, Yikang, HOI, Steven C. H., WANG, Xiaogang |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2018
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/4182 https://ink.library.smu.edu.sg/context/sis_research/article/5185/viewcontent/Question_GuidedHybridConvolution_2018_afv.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
Similar Items
-
REPRESENTATION LEARNING OF DATA WITH MULTIPLE MODALITIES WITH APPLICATIONS TO VISUAL QUESTION ANSWERING
by: ILIEVSKI ILIJA
Published: (2018) -
Dynamic fusion with intra-and inter-modality attention flow for visual question answering
by: GAO, Peng, et al.
Published: (2019) -
Aggregated community question answering
by: Snehasish Banerjee, et al.
Published: (2015) -
Exploring duality in visual question-driven top-down saliency
by: HE, Shengfeng, et al.
Published: (2020) -
Visual questioning and answering
by: Ong, Zavier Jian Le
Published: (2024)