Delving into inter-image invariance for unsupervised visual representations

Contrastive learning has recently shown immense potential in unsupervised visual representation learning. Existing studies in this track mainly focus on intra-image invariance learning. The learning typically uses rich intra-image transformations to construct positive pairs and then maximizes agreem...

Full description

Saved in:
Bibliographic Details
Main Authors: Xie, Jiahao, Zhan, Xiaohang, Liu, Ziwei, Ong, Yew-Soon, Loy, Chen Change
Other Authors: School of Computer Science and Engineering
Format: Article
Language:English
Published: 2023
Subjects:
Online Access:https://hdl.handle.net/10356/170424
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-170424
record_format dspace
spelling sg-ntu-dr.10356-1704242023-09-12T02:07:01Z Delving into inter-image invariance for unsupervised visual representations Xie, Jiahao Zhan, Xiaohang Liu, Ziwei Ong, Yew-Soon Loy, Chen Change School of Computer Science and Engineering Engineering::Computer science and engineering Unsupervised Learning Self-supervised Learning Contrastive learning has recently shown immense potential in unsupervised visual representation learning. Existing studies in this track mainly focus on intra-image invariance learning. The learning typically uses rich intra-image transformations to construct positive pairs and then maximizes agreement using a contrastive loss. The merits of inter-image invariance, conversely, remain much less explored. One major obstacle to exploit inter-image invariance is that it is unclear how to reliably construct inter-image positive pairs, and further derive effective supervision from them since no pair annotations are available. In this work, we present a comprehensive empirical study to better understand the role of inter-image invariance learning from three main constituting components: pseudo-label maintenance, sampling strategy, and decision boundary design. To facilitate the study, we introduce a unified and generic framework that supports the integration of unsupervised intra- and inter-image invariance learning. Through carefully-designed comparisons and analysis, multiple valuable observations are revealed: 1) online labels converge faster and perform better than offline labels; 2) semi-hard negative samples are more reliable and unbiased than hard negative samples; 3) a less stringent decision boundary is more favorable for inter-image invariance learning. With all the obtained recipes, our final model, namely InterCLR, shows consistent improvements over state-of-the-art intra-image invariance learning methods on multiple standard benchmarks. We hope this work will provide useful experience for devising effective unsupervised inter-image invariance learning. Code: https://github.com/open-mmlab/mmselfsup. Agency for Science, Technology and Research (A*STAR) Ministry of Education (MOE) This study is supported under the RIE2020 Industry Alignment Fund-Industry Collaboration Projects (IAF-ICP) Funding Initiative, as well as cash and in-kind contribution from the industry partner(s). The project is also supported by Singapore MOE AcRF Tier 2 (T2EP20120-0001), the Data Science and Artificial Intelligence Research Center at Nanyang Technological University. 2023-09-12T02:07:01Z 2023-09-12T02:07:01Z 2022 Journal Article Xie, J., Zhan, X., Liu, Z., Ong, Y. & Loy, C. C. (2022). Delving into inter-image invariance for unsupervised visual representations. International Journal of Computer Vision, 130(12), 2994-3013. https://dx.doi.org/10.1007/s11263-022-01681-x 0920-5691 https://hdl.handle.net/10356/170424 10.1007/s11263-022-01681-x 2-s2.0-85138719845 12 130 2994 3013 en T2EP20120-0001 International Journal of Computer Vision © 2022 The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature. All rights reserved.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering
Unsupervised Learning
Self-supervised Learning
spellingShingle Engineering::Computer science and engineering
Unsupervised Learning
Self-supervised Learning
Xie, Jiahao
Zhan, Xiaohang
Liu, Ziwei
Ong, Yew-Soon
Loy, Chen Change
Delving into inter-image invariance for unsupervised visual representations
description Contrastive learning has recently shown immense potential in unsupervised visual representation learning. Existing studies in this track mainly focus on intra-image invariance learning. The learning typically uses rich intra-image transformations to construct positive pairs and then maximizes agreement using a contrastive loss. The merits of inter-image invariance, conversely, remain much less explored. One major obstacle to exploit inter-image invariance is that it is unclear how to reliably construct inter-image positive pairs, and further derive effective supervision from them since no pair annotations are available. In this work, we present a comprehensive empirical study to better understand the role of inter-image invariance learning from three main constituting components: pseudo-label maintenance, sampling strategy, and decision boundary design. To facilitate the study, we introduce a unified and generic framework that supports the integration of unsupervised intra- and inter-image invariance learning. Through carefully-designed comparisons and analysis, multiple valuable observations are revealed: 1) online labels converge faster and perform better than offline labels; 2) semi-hard negative samples are more reliable and unbiased than hard negative samples; 3) a less stringent decision boundary is more favorable for inter-image invariance learning. With all the obtained recipes, our final model, namely InterCLR, shows consistent improvements over state-of-the-art intra-image invariance learning methods on multiple standard benchmarks. We hope this work will provide useful experience for devising effective unsupervised inter-image invariance learning. Code: https://github.com/open-mmlab/mmselfsup.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Xie, Jiahao
Zhan, Xiaohang
Liu, Ziwei
Ong, Yew-Soon
Loy, Chen Change
format Article
author Xie, Jiahao
Zhan, Xiaohang
Liu, Ziwei
Ong, Yew-Soon
Loy, Chen Change
author_sort Xie, Jiahao
title Delving into inter-image invariance for unsupervised visual representations
title_short Delving into inter-image invariance for unsupervised visual representations
title_full Delving into inter-image invariance for unsupervised visual representations
title_fullStr Delving into inter-image invariance for unsupervised visual representations
title_full_unstemmed Delving into inter-image invariance for unsupervised visual representations
title_sort delving into inter-image invariance for unsupervised visual representations
publishDate 2023
url https://hdl.handle.net/10356/170424
_version_ 1779156674192343040