A simple data mixing prior for improving self-supervised learning

Data mixing (e.g., Mixup, Cutmix, ResizeMix) is an essential component for advancing recognition models. In this paper, we focus on studying its effectiveness in the self-supervised setting. By noticing the mixed images that share the same source images are intrinsically related to each other, we he...

Full description

Saved in:
Bibliographic Details
Main Authors: REN, Sucheng, WANG, Huiyu, GAO, Zhengqi, HE, Shengfeng, YUILLE, Alan, ZHOU, Yuyin, XIE, Cihang
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8445
https://ink.library.smu.edu.sg/context/sis_research/article/9448/viewcontent/Ren_A_Simple_Data_Mixing_Prior_for_Improving_Self_Supervised_Learning_CVPR_2022_paper.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9448
record_format dspace
spelling sg-smu-ink.sis_research-94482024-01-04T09:54:09Z A simple data mixing prior for improving self-supervised learning REN, Sucheng WANG, Huiyu GAO, Zhengqi HE, Shengfeng YUILLE, Alan ZHOU, Yuyin XIE, Cihang Data mixing (e.g., Mixup, Cutmix, ResizeMix) is an essential component for advancing recognition models. In this paper, we focus on studying its effectiveness in the self-supervised setting. By noticing the mixed images that share the same source images are intrinsically related to each other, we hereby propose SDMP, short for Simple Data Mixing Prior, to capture this straightforward yet essential prior, and position such mixed images as additional positive pairs to facilitate self-supervised representation learning. Our experiments verify that the proposed SDMP enables data mixing to help a set of self-supervised learning frameworks (e.g., MoCo) achieve better accuracy and out-of-distribution robustness. More notably, our SDMP is the first method that successfully leverages data mixing to improve (rather than hurt) the performance of Vision Transformers in the self-supervised setting. Code is publicly available at https://github.com/OliverRensu/SDMP. 2022-06-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8445 info:doi/10.1109/CVPR52688.2022.01419 https://ink.library.smu.edu.sg/context/sis_research/article/9448/viewcontent/Ren_A_Simple_Data_Mixing_Prior_for_Improving_Self_Supervised_Learning_CVPR_2022_paper.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Categorization Representation learning Retrieval Self- & semi- & meta- recognition Detection Databases and Information Systems Graphics and Human Computer Interfaces
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Categorization
Representation learning
Retrieval
Self- & semi- & meta- recognition
Detection
Databases and Information Systems
Graphics and Human Computer Interfaces
spellingShingle Categorization
Representation learning
Retrieval
Self- & semi- & meta- recognition
Detection
Databases and Information Systems
Graphics and Human Computer Interfaces
REN, Sucheng
WANG, Huiyu
GAO, Zhengqi
HE, Shengfeng
YUILLE, Alan
ZHOU, Yuyin
XIE, Cihang
A simple data mixing prior for improving self-supervised learning
description Data mixing (e.g., Mixup, Cutmix, ResizeMix) is an essential component for advancing recognition models. In this paper, we focus on studying its effectiveness in the self-supervised setting. By noticing the mixed images that share the same source images are intrinsically related to each other, we hereby propose SDMP, short for Simple Data Mixing Prior, to capture this straightforward yet essential prior, and position such mixed images as additional positive pairs to facilitate self-supervised representation learning. Our experiments verify that the proposed SDMP enables data mixing to help a set of self-supervised learning frameworks (e.g., MoCo) achieve better accuracy and out-of-distribution robustness. More notably, our SDMP is the first method that successfully leverages data mixing to improve (rather than hurt) the performance of Vision Transformers in the self-supervised setting. Code is publicly available at https://github.com/OliverRensu/SDMP.
format text
author REN, Sucheng
WANG, Huiyu
GAO, Zhengqi
HE, Shengfeng
YUILLE, Alan
ZHOU, Yuyin
XIE, Cihang
author_facet REN, Sucheng
WANG, Huiyu
GAO, Zhengqi
HE, Shengfeng
YUILLE, Alan
ZHOU, Yuyin
XIE, Cihang
author_sort REN, Sucheng
title A simple data mixing prior for improving self-supervised learning
title_short A simple data mixing prior for improving self-supervised learning
title_full A simple data mixing prior for improving self-supervised learning
title_fullStr A simple data mixing prior for improving self-supervised learning
title_full_unstemmed A simple data mixing prior for improving self-supervised learning
title_sort simple data mixing prior for improving self-supervised learning
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/8445
https://ink.library.smu.edu.sg/context/sis_research/article/9448/viewcontent/Ren_A_Simple_Data_Mixing_Prior_for_Improving_Self_Supervised_Learning_CVPR_2022_paper.pdf
_version_ 1787590750744084480