Self-supervised spatio-temporal representation learning for videos by predicting motion and appearance statistics

We address the problem of video representation learning without human-annotated labels. While previous efforts address the problem by designing novel self-supervised tasks using video data, the learned features are merely on a frame-by-frame basis, which are not applicable to many video analytic tas...

Full description

Saved in:
Bibliographic Details
Main Authors: WANG, Jiangliu, JIAO, Jianbo, BAO, Linchao, HE, Shengfeng, LIU, Yunhui, LIU, Wei
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2019
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8439
https://ink.library.smu.edu.sg/context/sis_research/article/9442/viewcontent/Wang_Self_Supervised_Spatio_Temporal_Representation_Learning_for_Videos_by_Predicting_Motion_and_CVPR_2019_paper.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9442
record_format dspace
spelling sg-smu-ink.sis_research-94422024-01-04T09:56:22Z Self-supervised spatio-temporal representation learning for videos by predicting motion and appearance statistics WANG, Jiangliu JIAO, Jianbo BAO, Linchao HE, Shengfeng LIU, Yunhui LIU, Wei We address the problem of video representation learning without human-annotated labels. While previous efforts address the problem by designing novel self-supervised tasks using video data, the learned features are merely on a frame-by-frame basis, which are not applicable to many video analytic tasks where spatio-temporal features are prevailing. In this paper we propose a novel self-supervised approach to learn spatio-temporal features for video representation. Inspired by the success of two-stream approaches in video classification, we propose to learn visual features by regressing both motion and appearance statistics along spatial and temporal dimensions, given only the input video data. Specifically, we extract statistical concepts (fast-motion region and the corresponding dominant direction, spatio-temporal color diversity, dominant color, etc.) from simple patterns in both spatial and temporal domains. Unlike prior puzzles that are even hard for humans to solve, the proposed approach is consistent with human inherent visual habits and therefore easy to answer. We conduct extensive experiments with C3D to validate the effectiveness of our proposed approach. The experiments show that our approach can significantly improve the performance of C3D when applied to video classification tasks. Code is available at https://github.com/laura-wang/video repres mas. 2019-06-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8439 info:doi/10.1109/CVPR.2019.00413 https://ink.library.smu.edu.sg/context/sis_research/article/9442/viewcontent/Wang_Self_Supervised_Spatio_Temporal_Representation_Learning_for_Videos_by_Predicting_Motion_and_CVPR_2019_paper.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Representation learning Video analytics Artificial Intelligence and Robotics Databases and Information Systems
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Representation learning
Video analytics
Artificial Intelligence and Robotics
Databases and Information Systems
spellingShingle Representation learning
Video analytics
Artificial Intelligence and Robotics
Databases and Information Systems
WANG, Jiangliu
JIAO, Jianbo
BAO, Linchao
HE, Shengfeng
LIU, Yunhui
LIU, Wei
Self-supervised spatio-temporal representation learning for videos by predicting motion and appearance statistics
description We address the problem of video representation learning without human-annotated labels. While previous efforts address the problem by designing novel self-supervised tasks using video data, the learned features are merely on a frame-by-frame basis, which are not applicable to many video analytic tasks where spatio-temporal features are prevailing. In this paper we propose a novel self-supervised approach to learn spatio-temporal features for video representation. Inspired by the success of two-stream approaches in video classification, we propose to learn visual features by regressing both motion and appearance statistics along spatial and temporal dimensions, given only the input video data. Specifically, we extract statistical concepts (fast-motion region and the corresponding dominant direction, spatio-temporal color diversity, dominant color, etc.) from simple patterns in both spatial and temporal domains. Unlike prior puzzles that are even hard for humans to solve, the proposed approach is consistent with human inherent visual habits and therefore easy to answer. We conduct extensive experiments with C3D to validate the effectiveness of our proposed approach. The experiments show that our approach can significantly improve the performance of C3D when applied to video classification tasks. Code is available at https://github.com/laura-wang/video repres mas.
format text
author WANG, Jiangliu
JIAO, Jianbo
BAO, Linchao
HE, Shengfeng
LIU, Yunhui
LIU, Wei
author_facet WANG, Jiangliu
JIAO, Jianbo
BAO, Linchao
HE, Shengfeng
LIU, Yunhui
LIU, Wei
author_sort WANG, Jiangliu
title Self-supervised spatio-temporal representation learning for videos by predicting motion and appearance statistics
title_short Self-supervised spatio-temporal representation learning for videos by predicting motion and appearance statistics
title_full Self-supervised spatio-temporal representation learning for videos by predicting motion and appearance statistics
title_fullStr Self-supervised spatio-temporal representation learning for videos by predicting motion and appearance statistics
title_full_unstemmed Self-supervised spatio-temporal representation learning for videos by predicting motion and appearance statistics
title_sort self-supervised spatio-temporal representation learning for videos by predicting motion and appearance statistics
publisher Institutional Knowledge at Singapore Management University
publishDate 2019
url https://ink.library.smu.edu.sg/sis_research/8439
https://ink.library.smu.edu.sg/context/sis_research/article/9442/viewcontent/Wang_Self_Supervised_Spatio_Temporal_Representation_Learning_for_Videos_by_Predicting_Motion_and_CVPR_2019_paper.pdf
_version_ 1787590749696557056