Explore the influential samples in domain generalization
Domain Generalization (DG) aims to learn a model that generalizes in testing domains unseen from training. All DG methods assume that the domain-invariant features can be learned by discarding the domain-specific ones. However, in practice, the learned invariant features usually contain "spurio...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Master by Research |
Language: | English |
Published: |
Nanyang Technological University
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/172753 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-172753 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1727532024-01-04T06:32:51Z Explore the influential samples in domain generalization Wu, Zike Hanwang Zhang School of Computer Science and Engineering hanwangzhang@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision Domain Generalization (DG) aims to learn a model that generalizes in testing domains unseen from training. All DG methods assume that the domain-invariant features can be learned by discarding the domain-specific ones. However, in practice, the learned invariant features usually contain "spurious invariance" that is only invariant across training domains but still variant to testing ones. We point out that this is because the contribution of the minority training samples without such spurious invariance is outgunned. Therefore, we are motivated to split these samples out of the original domains to form a new one, to which the spurious invariance is no longer invariant and thus removed. We present a cross-domain influence-based method, Domain+, to obtain the new domain. Specifically, for each sample per training domain, we estimate its influence by up-weighting it and then calculating how much the invariance loss of the other training domains changes---the more it changes, the higher the influence, and the more likely the sample belongs to the new domain. Then, with the split domains, we can deploy any off-the-shelf DG methods to achieve better generalization. We benchmark Domain+ on DomainBed and show that it helps existing SOTA methods achieve new SOTAs. Master of Engineering 2023-12-19T08:07:54Z 2023-12-19T08:07:54Z 2023 Thesis-Master by Research Wu, Z. (2023). Explore the influential samples in domain generalization. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/172753 https://hdl.handle.net/10356/172753 10.32657/10356/172753 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision |
spellingShingle |
Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision Wu, Zike Explore the influential samples in domain generalization |
description |
Domain Generalization (DG) aims to learn a model that generalizes in testing domains unseen from training. All DG methods assume that the domain-invariant features can be learned by discarding the domain-specific ones. However, in practice, the learned invariant features usually contain "spurious invariance" that is only invariant across training domains but still variant to testing ones. We point out that this is because the contribution of the minority training samples without such spurious invariance is outgunned. Therefore, we are motivated to split these samples out of the original domains to form a new one, to which the spurious invariance is no longer invariant and thus removed. We present a cross-domain influence-based method, Domain+, to obtain the new domain. Specifically, for each sample per training domain, we estimate its influence by up-weighting it and then calculating how much the invariance loss of the other training domains changes---the more it changes, the higher the influence, and the more likely the sample belongs to the new domain. Then, with the split domains, we can deploy any off-the-shelf DG methods to achieve better generalization. We benchmark Domain+ on DomainBed and show that it helps existing SOTA methods achieve new SOTAs. |
author2 |
Hanwang Zhang |
author_facet |
Hanwang Zhang Wu, Zike |
format |
Thesis-Master by Research |
author |
Wu, Zike |
author_sort |
Wu, Zike |
title |
Explore the influential samples in domain generalization |
title_short |
Explore the influential samples in domain generalization |
title_full |
Explore the influential samples in domain generalization |
title_fullStr |
Explore the influential samples in domain generalization |
title_full_unstemmed |
Explore the influential samples in domain generalization |
title_sort |
explore the influential samples in domain generalization |
publisher |
Nanyang Technological University |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/172753 |
_version_ |
1787590724066213888 |