Constructing holistic spatio-temporal scene graph for video semantic role labeling
As one of the core video semantic understanding tasks, Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL,...
Saved in:
Main Authors: | , , , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2023
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8290 https://ink.library.smu.edu.sg/context/sis_research/article/9293/viewcontent/2308.05081.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9293 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-92932023-12-20T03:02:37Z Constructing holistic spatio-temporal scene graph for video semantic role labeling ZHAO, Yu FEI, Hao CAO, Yixin LI, Bobo ZHANG, Meishan WEI, Jianguo ZHANG, Min CHUA, Tat-Seng As one of the core video semantic understanding tasks, Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, e.g., filtering noisy branches and newly building informative connections, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods. Our HostSG representation shows greater potential to facilitate a broader range of other video understanding tasks. 2023-11-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8290 info:doi/10.1145/3581783.3612096 https://ink.library.smu.edu.sg/context/sis_research/article/9293/viewcontent/2308.05081.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University video understanding semantics role labeling event extraction scene graph Graphics and Human Computer Interfaces Numerical Analysis and Scientific Computing |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
video understanding semantics role labeling event extraction scene graph Graphics and Human Computer Interfaces Numerical Analysis and Scientific Computing |
spellingShingle |
video understanding semantics role labeling event extraction scene graph Graphics and Human Computer Interfaces Numerical Analysis and Scientific Computing ZHAO, Yu FEI, Hao CAO, Yixin LI, Bobo ZHANG, Meishan WEI, Jianguo ZHANG, Min CHUA, Tat-Seng Constructing holistic spatio-temporal scene graph for video semantic role labeling |
description |
As one of the core video semantic understanding tasks, Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, e.g., filtering noisy branches and newly building informative connections, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods. Our HostSG representation shows greater potential to facilitate a broader range of other video understanding tasks. |
format |
text |
author |
ZHAO, Yu FEI, Hao CAO, Yixin LI, Bobo ZHANG, Meishan WEI, Jianguo ZHANG, Min CHUA, Tat-Seng |
author_facet |
ZHAO, Yu FEI, Hao CAO, Yixin LI, Bobo ZHANG, Meishan WEI, Jianguo ZHANG, Min CHUA, Tat-Seng |
author_sort |
ZHAO, Yu |
title |
Constructing holistic spatio-temporal scene graph for video semantic role labeling |
title_short |
Constructing holistic spatio-temporal scene graph for video semantic role labeling |
title_full |
Constructing holistic spatio-temporal scene graph for video semantic role labeling |
title_fullStr |
Constructing holistic spatio-temporal scene graph for video semantic role labeling |
title_full_unstemmed |
Constructing holistic spatio-temporal scene graph for video semantic role labeling |
title_sort |
constructing holistic spatio-temporal scene graph for video semantic role labeling |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2023 |
url |
https://ink.library.smu.edu.sg/sis_research/8290 https://ink.library.smu.edu.sg/context/sis_research/article/9293/viewcontent/2308.05081.pdf |
_version_ |
1787136836407132160 |