Learning language to symbol and language to vision mapping for visual grounding
Visual Grounding (VG) is a task of locating a specific object in an image semantically matching a given linguistic expression. The mapping of the linguistic and visual contents and the understanding of diverse linguistic expressions are the two challenges of this task. The performance of visual grou...
Saved in:
Main Authors: | He, Su, Yang, Xiaofeng, Lin, Guosheng |
---|---|
Other Authors: | School of Computer Science and Engineering |
Format: | Article |
Language: | English |
Published: |
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/161552 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
Similar Items
-
Is a high tone pointy? Speakers of different languages match Mandarin Chinese tones to visual shapes differently
by: Shang, Nan, et al.
Published: (2018) -
Enhancing visual grounding in vision-language pre-training with position-guided text prompts
by: WANG, Alex Jinpeng, et al.
Published: (2024) -
Neural logic vision language explainer
by: Yang, Xiaofeng, et al.
Published: (2023) -
Demo abstract: VGGlass - Demonstrating visual grounding and localization synergy with a LiDAR-enabled smart-glass
by: RATHNAYAKE, Darshana, et al.
Published: (2023) -
Vision language representation learning
by: Yang, Xiaofeng
Published: (2023)