Delving into inter-image invariance for unsupervised visual representations
Contrastive learning has recently shown immense potential in unsupervised visual representation learning. Existing studies in this track mainly focus on intra-image invariance learning. The learning typically uses rich intra-image transformations to construct positive pairs and then maximizes agreem...
Saved in:
Main Authors: | , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/170424 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-170424 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1704242023-09-12T02:07:01Z Delving into inter-image invariance for unsupervised visual representations Xie, Jiahao Zhan, Xiaohang Liu, Ziwei Ong, Yew-Soon Loy, Chen Change School of Computer Science and Engineering Engineering::Computer science and engineering Unsupervised Learning Self-supervised Learning Contrastive learning has recently shown immense potential in unsupervised visual representation learning. Existing studies in this track mainly focus on intra-image invariance learning. The learning typically uses rich intra-image transformations to construct positive pairs and then maximizes agreement using a contrastive loss. The merits of inter-image invariance, conversely, remain much less explored. One major obstacle to exploit inter-image invariance is that it is unclear how to reliably construct inter-image positive pairs, and further derive effective supervision from them since no pair annotations are available. In this work, we present a comprehensive empirical study to better understand the role of inter-image invariance learning from three main constituting components: pseudo-label maintenance, sampling strategy, and decision boundary design. To facilitate the study, we introduce a unified and generic framework that supports the integration of unsupervised intra- and inter-image invariance learning. Through carefully-designed comparisons and analysis, multiple valuable observations are revealed: 1) online labels converge faster and perform better than offline labels; 2) semi-hard negative samples are more reliable and unbiased than hard negative samples; 3) a less stringent decision boundary is more favorable for inter-image invariance learning. With all the obtained recipes, our final model, namely InterCLR, shows consistent improvements over state-of-the-art intra-image invariance learning methods on multiple standard benchmarks. We hope this work will provide useful experience for devising effective unsupervised inter-image invariance learning. Code: https://github.com/open-mmlab/mmselfsup. Agency for Science, Technology and Research (A*STAR) Ministry of Education (MOE) This study is supported under the RIE2020 Industry Alignment Fund-Industry Collaboration Projects (IAF-ICP) Funding Initiative, as well as cash and in-kind contribution from the industry partner(s). The project is also supported by Singapore MOE AcRF Tier 2 (T2EP20120-0001), the Data Science and Artificial Intelligence Research Center at Nanyang Technological University. 2023-09-12T02:07:01Z 2023-09-12T02:07:01Z 2022 Journal Article Xie, J., Zhan, X., Liu, Z., Ong, Y. & Loy, C. C. (2022). Delving into inter-image invariance for unsupervised visual representations. International Journal of Computer Vision, 130(12), 2994-3013. https://dx.doi.org/10.1007/s11263-022-01681-x 0920-5691 https://hdl.handle.net/10356/170424 10.1007/s11263-022-01681-x 2-s2.0-85138719845 12 130 2994 3013 en T2EP20120-0001 International Journal of Computer Vision © 2022 The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Unsupervised Learning Self-supervised Learning |
spellingShingle |
Engineering::Computer science and engineering Unsupervised Learning Self-supervised Learning Xie, Jiahao Zhan, Xiaohang Liu, Ziwei Ong, Yew-Soon Loy, Chen Change Delving into inter-image invariance for unsupervised visual representations |
description |
Contrastive learning has recently shown immense potential in unsupervised visual representation learning. Existing studies in this track mainly focus on intra-image invariance learning. The learning typically uses rich intra-image transformations to construct positive pairs and then maximizes agreement using a contrastive loss. The merits of inter-image invariance, conversely, remain much less explored. One major obstacle to exploit inter-image invariance is that it is unclear how to reliably construct inter-image positive pairs, and further derive effective supervision from them since no pair annotations are available. In this work, we present a comprehensive empirical study to better understand the role of inter-image invariance learning from three main constituting components: pseudo-label maintenance, sampling strategy, and decision boundary design. To facilitate the study, we introduce a unified and generic framework that supports the integration of unsupervised intra- and inter-image invariance learning. Through carefully-designed comparisons and analysis, multiple valuable observations are revealed: 1) online labels converge faster and perform better than offline labels; 2) semi-hard negative samples are more reliable and unbiased than hard negative samples; 3) a less stringent decision boundary is more favorable for inter-image invariance learning. With all the obtained recipes, our final model, namely InterCLR, shows consistent improvements over state-of-the-art intra-image invariance learning methods on multiple standard benchmarks. We hope this work will provide useful experience for devising effective unsupervised inter-image invariance learning. Code: https://github.com/open-mmlab/mmselfsup. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Xie, Jiahao Zhan, Xiaohang Liu, Ziwei Ong, Yew-Soon Loy, Chen Change |
format |
Article |
author |
Xie, Jiahao Zhan, Xiaohang Liu, Ziwei Ong, Yew-Soon Loy, Chen Change |
author_sort |
Xie, Jiahao |
title |
Delving into inter-image invariance for unsupervised visual representations |
title_short |
Delving into inter-image invariance for unsupervised visual representations |
title_full |
Delving into inter-image invariance for unsupervised visual representations |
title_fullStr |
Delving into inter-image invariance for unsupervised visual representations |
title_full_unstemmed |
Delving into inter-image invariance for unsupervised visual representations |
title_sort |
delving into inter-image invariance for unsupervised visual representations |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/170424 |
_version_ |
1779156674192343040 |