Delving deep into many-to-many attention for few-shot video object segmentation
This paper tackles the task of Few-Shot Video Object Segmentation (FSVOS), i.e., segmenting objects in the query videos with certain class specified in a few labeled support images. The key is to model the relationship between the query videos and the support images for propagating the object inform...
Saved in:
Main Authors: | , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2021
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8527 https://ink.library.smu.edu.sg/context/sis_research/article/9530/viewcontent/Delving_deep_into_many_to_many_attention_for_few_shot_video_object_segmentation.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9530 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-95302024-01-22T15:00:08Z Delving deep into many-to-many attention for few-shot video object segmentation CHEN, Haoxin WU, Hanjie ZHAO, Nanxuan REN, Sucheng HE, Shengfeng This paper tackles the task of Few-Shot Video Object Segmentation (FSVOS), i.e., segmenting objects in the query videos with certain class specified in a few labeled support images. The key is to model the relationship between the query videos and the support images for propagating the object information. This is a many-to-many problem and often relies on full-rank attention, which is computationally intensive. In this paper, we propose a novel Domain Agent Network (DAN), breaking down the full-rank attention into two smaller ones. We consider one single frame of the query video as the domain agent, bridging between the support images and the query video. Our DAN allows a linear space and time complexity as opposed to the original quadratic form with no loss of performance. In addition, we introduce a learning strategy by combining meta-learning with online learning to further improve the segmentation accuracy. We build a FSVOS benchmark on the Youtube-VIS dataset and conduct experiments to demonstrate that our method outperforms baselines on both computational cost and accuracy, achieving the state-of-the-art performance 2021-06-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8527 info:doi/10.1109/CVPR46437.2021.01382 https://ink.library.smu.edu.sg/context/sis_research/article/9530/viewcontent/Delving_deep_into_many_to_many_attention_for_few_shot_video_object_segmentation.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Agent network; Breakings; Linear spaces; Linear time; Many to many; Novel domain; Object information; Query video; Single frames; Video objects segmentations Graphics and Human Computer Interfaces Numerical Analysis and Scientific Computing |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Agent network; Breakings; Linear spaces; Linear time; Many to many; Novel domain; Object information; Query video; Single frames; Video objects segmentations Graphics and Human Computer Interfaces Numerical Analysis and Scientific Computing |
spellingShingle |
Agent network; Breakings; Linear spaces; Linear time; Many to many; Novel domain; Object information; Query video; Single frames; Video objects segmentations Graphics and Human Computer Interfaces Numerical Analysis and Scientific Computing CHEN, Haoxin WU, Hanjie ZHAO, Nanxuan REN, Sucheng HE, Shengfeng Delving deep into many-to-many attention for few-shot video object segmentation |
description |
This paper tackles the task of Few-Shot Video Object Segmentation (FSVOS), i.e., segmenting objects in the query videos with certain class specified in a few labeled support images. The key is to model the relationship between the query videos and the support images for propagating the object information. This is a many-to-many problem and often relies on full-rank attention, which is computationally intensive. In this paper, we propose a novel Domain Agent Network (DAN), breaking down the full-rank attention into two smaller ones. We consider one single frame of the query video as the domain agent, bridging between the support images and the query video. Our DAN allows a linear space and time complexity as opposed to the original quadratic form with no loss of performance. In addition, we introduce a learning strategy by combining meta-learning with online learning to further improve the segmentation accuracy. We build a FSVOS benchmark on the Youtube-VIS dataset and conduct experiments to demonstrate that our method outperforms baselines on both computational cost and accuracy, achieving the state-of-the-art performance |
format |
text |
author |
CHEN, Haoxin WU, Hanjie ZHAO, Nanxuan REN, Sucheng HE, Shengfeng |
author_facet |
CHEN, Haoxin WU, Hanjie ZHAO, Nanxuan REN, Sucheng HE, Shengfeng |
author_sort |
CHEN, Haoxin |
title |
Delving deep into many-to-many attention for few-shot video object segmentation |
title_short |
Delving deep into many-to-many attention for few-shot video object segmentation |
title_full |
Delving deep into many-to-many attention for few-shot video object segmentation |
title_fullStr |
Delving deep into many-to-many attention for few-shot video object segmentation |
title_full_unstemmed |
Delving deep into many-to-many attention for few-shot video object segmentation |
title_sort |
delving deep into many-to-many attention for few-shot video object segmentation |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2021 |
url |
https://ink.library.smu.edu.sg/sis_research/8527 https://ink.library.smu.edu.sg/context/sis_research/article/9530/viewcontent/Delving_deep_into_many_to_many_attention_for_few_shot_video_object_segmentation.pdf |
_version_ |
1789483259223080960 |