A simple data mixing prior for improving self-supervised learning
Data mixing (e.g., Mixup, Cutmix, ResizeMix) is an essential component for advancing recognition models. In this paper, we focus on studying its effectiveness in the self-supervised setting. By noticing the mixed images that share the same source images are intrinsically related to each other, we he...
Saved in:
Main Authors: | , , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2022
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8445 https://ink.library.smu.edu.sg/context/sis_research/article/9448/viewcontent/Ren_A_Simple_Data_Mixing_Prior_for_Improving_Self_Supervised_Learning_CVPR_2022_paper.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9448 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-94482024-01-04T09:54:09Z A simple data mixing prior for improving self-supervised learning REN, Sucheng WANG, Huiyu GAO, Zhengqi HE, Shengfeng YUILLE, Alan ZHOU, Yuyin XIE, Cihang Data mixing (e.g., Mixup, Cutmix, ResizeMix) is an essential component for advancing recognition models. In this paper, we focus on studying its effectiveness in the self-supervised setting. By noticing the mixed images that share the same source images are intrinsically related to each other, we hereby propose SDMP, short for Simple Data Mixing Prior, to capture this straightforward yet essential prior, and position such mixed images as additional positive pairs to facilitate self-supervised representation learning. Our experiments verify that the proposed SDMP enables data mixing to help a set of self-supervised learning frameworks (e.g., MoCo) achieve better accuracy and out-of-distribution robustness. More notably, our SDMP is the first method that successfully leverages data mixing to improve (rather than hurt) the performance of Vision Transformers in the self-supervised setting. Code is publicly available at https://github.com/OliverRensu/SDMP. 2022-06-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8445 info:doi/10.1109/CVPR52688.2022.01419 https://ink.library.smu.edu.sg/context/sis_research/article/9448/viewcontent/Ren_A_Simple_Data_Mixing_Prior_for_Improving_Self_Supervised_Learning_CVPR_2022_paper.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Categorization Representation learning Retrieval Self- & semi- & meta- recognition Detection Databases and Information Systems Graphics and Human Computer Interfaces |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Categorization Representation learning Retrieval Self- & semi- & meta- recognition Detection Databases and Information Systems Graphics and Human Computer Interfaces |
spellingShingle |
Categorization Representation learning Retrieval Self- & semi- & meta- recognition Detection Databases and Information Systems Graphics and Human Computer Interfaces REN, Sucheng WANG, Huiyu GAO, Zhengqi HE, Shengfeng YUILLE, Alan ZHOU, Yuyin XIE, Cihang A simple data mixing prior for improving self-supervised learning |
description |
Data mixing (e.g., Mixup, Cutmix, ResizeMix) is an essential component for advancing recognition models. In this paper, we focus on studying its effectiveness in the self-supervised setting. By noticing the mixed images that share the same source images are intrinsically related to each other, we hereby propose SDMP, short for Simple Data Mixing Prior, to capture this straightforward yet essential prior, and position such mixed images as additional positive pairs to facilitate self-supervised representation learning. Our experiments verify that the proposed SDMP enables data mixing to help a set of self-supervised learning frameworks (e.g., MoCo) achieve better accuracy and out-of-distribution robustness. More notably, our SDMP is the first method that successfully leverages data mixing to improve (rather than hurt) the performance of Vision Transformers in the self-supervised setting. Code is publicly available at https://github.com/OliverRensu/SDMP. |
format |
text |
author |
REN, Sucheng WANG, Huiyu GAO, Zhengqi HE, Shengfeng YUILLE, Alan ZHOU, Yuyin XIE, Cihang |
author_facet |
REN, Sucheng WANG, Huiyu GAO, Zhengqi HE, Shengfeng YUILLE, Alan ZHOU, Yuyin XIE, Cihang |
author_sort |
REN, Sucheng |
title |
A simple data mixing prior for improving self-supervised learning |
title_short |
A simple data mixing prior for improving self-supervised learning |
title_full |
A simple data mixing prior for improving self-supervised learning |
title_fullStr |
A simple data mixing prior for improving self-supervised learning |
title_full_unstemmed |
A simple data mixing prior for improving self-supervised learning |
title_sort |
simple data mixing prior for improving self-supervised learning |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2022 |
url |
https://ink.library.smu.edu.sg/sis_research/8445 https://ink.library.smu.edu.sg/context/sis_research/article/9448/viewcontent/Ren_A_Simple_Data_Mixing_Prior_for_Improving_Self_Supervised_Learning_CVPR_2022_paper.pdf |
_version_ |
1787590750744084480 |