Equivariance and invariance inductive bias for learning from insufficient data

We are interested in learning robust models from insufficient data, without the need for any externally pre-trained model checkpoints. First, compared to sufficient data, we show why insufficient data renders the model more easily biased to the limited training environments that are usually differen...

Full description

Saved in:
Bibliographic Details
Main Authors: WANG, Tan, SUN, Qianru, PRANATA, Sugiri, JAYASHREE, Karlekar, ZHANG, Hanwang
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7513
https://ink.library.smu.edu.sg/context/sis_research/article/8516/viewcontent/ECCV2022_Vipriors_WangTan.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8516
record_format dspace
spelling sg-smu-ink.sis_research-85162023-08-07T00:52:16Z Equivariance and invariance inductive bias for learning from insufficient data WANG, Tan SUN, Qianru PRANATA, Sugiri JAYASHREE, Karlekar ZHANG, Hanwang We are interested in learning robust models from insufficient data, without the need for any externally pre-trained model checkpoints. First, compared to sufficient data, we show why insufficient data renders the model more easily biased to the limited training environments that are usually different from testing. For example, if all the training "swan" samples are "white", the model may wrongly use the "white" environment to represent the intrinsic class "swan". Then, we justify that equivariance inductive bias can retain the class feature while invariance inductive bias can remove the environmental feature, leaving only the class feature that generalizes to any testing environmental changes. To impose them on learning, for equivariance, we demonstrate that any off-the-shelf contrastive-based self-supervised feature learning method can be deployed; for invariance, we propose a class-wise invariant risk minimization (IRM) that efficiently tackles the challenge of missing environmental annotation in conventional IRM. State-of-the-art experimental results on real-world visual benchmarks (NICO and VIPriors ImageNet) validate the great potential of the two inductive biases in reducing training data and parameters significantly. 2022-10-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7513 info:doi/10.1007/978-3-031-20083-0_15 https://ink.library.smu.edu.sg/context/sis_research/article/8516/viewcontent/ECCV2022_Vipriors_WangTan.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Inductive Bias Equivariance Invariant Risk Minimization Databases and Information Systems Graphics and Human Computer Interfaces Numerical Analysis and Scientific Computing
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Inductive Bias
Equivariance
Invariant Risk Minimization
Databases and Information Systems
Graphics and Human Computer Interfaces
Numerical Analysis and Scientific Computing
spellingShingle Inductive Bias
Equivariance
Invariant Risk Minimization
Databases and Information Systems
Graphics and Human Computer Interfaces
Numerical Analysis and Scientific Computing
WANG, Tan
SUN, Qianru
PRANATA, Sugiri
JAYASHREE, Karlekar
ZHANG, Hanwang
Equivariance and invariance inductive bias for learning from insufficient data
description We are interested in learning robust models from insufficient data, without the need for any externally pre-trained model checkpoints. First, compared to sufficient data, we show why insufficient data renders the model more easily biased to the limited training environments that are usually different from testing. For example, if all the training "swan" samples are "white", the model may wrongly use the "white" environment to represent the intrinsic class "swan". Then, we justify that equivariance inductive bias can retain the class feature while invariance inductive bias can remove the environmental feature, leaving only the class feature that generalizes to any testing environmental changes. To impose them on learning, for equivariance, we demonstrate that any off-the-shelf contrastive-based self-supervised feature learning method can be deployed; for invariance, we propose a class-wise invariant risk minimization (IRM) that efficiently tackles the challenge of missing environmental annotation in conventional IRM. State-of-the-art experimental results on real-world visual benchmarks (NICO and VIPriors ImageNet) validate the great potential of the two inductive biases in reducing training data and parameters significantly.
format text
author WANG, Tan
SUN, Qianru
PRANATA, Sugiri
JAYASHREE, Karlekar
ZHANG, Hanwang
author_facet WANG, Tan
SUN, Qianru
PRANATA, Sugiri
JAYASHREE, Karlekar
ZHANG, Hanwang
author_sort WANG, Tan
title Equivariance and invariance inductive bias for learning from insufficient data
title_short Equivariance and invariance inductive bias for learning from insufficient data
title_full Equivariance and invariance inductive bias for learning from insufficient data
title_fullStr Equivariance and invariance inductive bias for learning from insufficient data
title_full_unstemmed Equivariance and invariance inductive bias for learning from insufficient data
title_sort equivariance and invariance inductive bias for learning from insufficient data
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/7513
https://ink.library.smu.edu.sg/context/sis_research/article/8516/viewcontent/ECCV2022_Vipriors_WangTan.pdf
_version_ 1773551433751199744