Reciprocal transformations for unsupervised video object segmentation

Unsupervised video object segmentation (UVOS) aims at segmenting the primary objects in videos without any human intervention. Due to the lack of prior knowledge about the primary objects, identifying them from videos is the major challenge of UVOS. Previous methods often regard the moving objects a...

Full description

Saved in:
Bibliographic Details
Main Authors: REN, Sucheng, LIU, Wenxi, LIU, Yongtuo, CHEN, Haoxin, HAN, Guoqiang, HE, Shengfeng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2021
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8441
https://ink.library.smu.edu.sg/context/sis_research/article/9444/viewcontent/Ren_Reciprocal_Transformations_for_Unsupervised_Video_Object_Segmentation_CVPR_2021_paper.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9444
record_format dspace
spelling sg-smu-ink.sis_research-94442024-01-04T09:55:33Z Reciprocal transformations for unsupervised video object segmentation REN, Sucheng LIU, Wenxi LIU, Yongtuo CHEN, Haoxin HAN, Guoqiang HE, Shengfeng Unsupervised video object segmentation (UVOS) aims at segmenting the primary objects in videos without any human intervention. Due to the lack of prior knowledge about the primary objects, identifying them from videos is the major challenge of UVOS. Previous methods often regard the moving objects as primary ones and rely on optical flow to capture the motion cues in videos, but the flow information alone is insufficient to distinguish the primary objects from the background objects that move together. This is because, when the noisy motion features are combined with the appearance features, the localization of the primary objects is misguided. To address this problem, we propose a novel reciprocal transformation network to discover primary objects by correlating three key factors: the intra-frame contrast, the motion cues, and temporal coherence of recurring objects. Each corresponds to a representative type of primary object, and our reciprocal mechanism enables an organic coordination of them to effectively remove ambiguous distractions from videos. Additionally, to exclude the information of the moving background objects from motion features, our transformation module enables to reciprocally transform the appearance features to enhance the motion features, so as to focus on the moving objects with salient appearance while removing the co-moving outliers. Experiments on the public benchmarks demonstrate that our model significantly outperforms the state-of-the-art methods. Code is available at https://github.com/OliverRensu/RTNet. 2021-06-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8441 info:doi/10.1109/CVPR46437.2021.01520 https://ink.library.smu.edu.sg/context/sis_research/article/9444/viewcontent/Ren_Reciprocal_Transformations_for_Unsupervised_Video_Object_Segmentation_CVPR_2021_paper.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Computer vision Image segmentation Background objects Flow informations Human intervention Key factors Localisation Motion cues Motion features Moving objects Prior-knowledge Video objects segmentations Databases and Information Systems Graphics and Human Computer Interfaces
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Computer vision
Image segmentation
Background objects
Flow informations
Human intervention
Key factors
Localisation
Motion cues
Motion features
Moving objects
Prior-knowledge
Video objects segmentations
Databases and Information Systems
Graphics and Human Computer Interfaces
spellingShingle Computer vision
Image segmentation
Background objects
Flow informations
Human intervention
Key factors
Localisation
Motion cues
Motion features
Moving objects
Prior-knowledge
Video objects segmentations
Databases and Information Systems
Graphics and Human Computer Interfaces
REN, Sucheng
LIU, Wenxi
LIU, Yongtuo
CHEN, Haoxin
HAN, Guoqiang
HE, Shengfeng
Reciprocal transformations for unsupervised video object segmentation
description Unsupervised video object segmentation (UVOS) aims at segmenting the primary objects in videos without any human intervention. Due to the lack of prior knowledge about the primary objects, identifying them from videos is the major challenge of UVOS. Previous methods often regard the moving objects as primary ones and rely on optical flow to capture the motion cues in videos, but the flow information alone is insufficient to distinguish the primary objects from the background objects that move together. This is because, when the noisy motion features are combined with the appearance features, the localization of the primary objects is misguided. To address this problem, we propose a novel reciprocal transformation network to discover primary objects by correlating three key factors: the intra-frame contrast, the motion cues, and temporal coherence of recurring objects. Each corresponds to a representative type of primary object, and our reciprocal mechanism enables an organic coordination of them to effectively remove ambiguous distractions from videos. Additionally, to exclude the information of the moving background objects from motion features, our transformation module enables to reciprocally transform the appearance features to enhance the motion features, so as to focus on the moving objects with salient appearance while removing the co-moving outliers. Experiments on the public benchmarks demonstrate that our model significantly outperforms the state-of-the-art methods. Code is available at https://github.com/OliverRensu/RTNet.
format text
author REN, Sucheng
LIU, Wenxi
LIU, Yongtuo
CHEN, Haoxin
HAN, Guoqiang
HE, Shengfeng
author_facet REN, Sucheng
LIU, Wenxi
LIU, Yongtuo
CHEN, Haoxin
HAN, Guoqiang
HE, Shengfeng
author_sort REN, Sucheng
title Reciprocal transformations for unsupervised video object segmentation
title_short Reciprocal transformations for unsupervised video object segmentation
title_full Reciprocal transformations for unsupervised video object segmentation
title_fullStr Reciprocal transformations for unsupervised video object segmentation
title_full_unstemmed Reciprocal transformations for unsupervised video object segmentation
title_sort reciprocal transformations for unsupervised video object segmentation
publisher Institutional Knowledge at Singapore Management University
publishDate 2021
url https://ink.library.smu.edu.sg/sis_research/8441
https://ink.library.smu.edu.sg/context/sis_research/article/9444/viewcontent/Ren_Reciprocal_Transformations_for_Unsupervised_Video_Object_Segmentation_CVPR_2021_paper.pdf
_version_ 1787590750050975744