Unsupervised domain adaptation via importance sampling

Unsupervised domain adaptation aims to generalize a model from the label-rich source domain to the unlabeled target domain. Existing works mainly focus on aligning the global distribution statistics between source and target domains. However, they neglect distractions from the unexpected noisy sampl...

Full description

Saved in:
Bibliographic Details
Main Authors: XU, Xuemiao, HE, Hai, ZHANG, Huaidong, XU, Yangyang, HE, Shengfeng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2020
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7884
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8887
record_format dspace
spelling sg-smu-ink.sis_research-88872023-06-15T09:00:05Z Unsupervised domain adaptation via importance sampling XU, Xuemiao HE, Hai ZHANG, Huaidong XU, Yangyang HE, Shengfeng Unsupervised domain adaptation aims to generalize a model from the label-rich source domain to the unlabeled target domain. Existing works mainly focus on aligning the global distribution statistics between source and target domains. However, they neglect distractions from the unexpected noisy samples in domain distribution estimation, leading to domain misalignment or even negative transfer. In this paper, we present an importance sampling method for domain adaptation (ISDA), to measure sample contributions according to their "informative" levels. In particular, informative samples, as well as outliers, can be effectively modeled using feature-norm and prediction entropy of the network. The importance of information is further formulated as the importance sampling losses in features and label spaces. In this way, the proposed model mitigates the noisy outliers while enhancing the important samples during domain alignment. In addition, our model is easy to implement yet effective, and it does not introduce any extra parameters. Extensive experiments on several benchmark datasets show that our method outperforms state-of-the-art methods under both the standard and partial domain adaptation settings. 2020-12-01T08:00:00Z text https://ink.library.smu.edu.sg/sis_research/7884 info:doi/10.1109/TCSVT.2019.2963318 Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Feature extraction Entropy Tuning Estimation Monte Carlo methods Adaptation models Noise measurement Domain adaptation deep learning distribution sampling Information Security
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Feature extraction
Entropy
Tuning
Estimation
Monte Carlo methods
Adaptation models
Noise measurement
Domain adaptation
deep learning
distribution sampling
Information Security
spellingShingle Feature extraction
Entropy
Tuning
Estimation
Monte Carlo methods
Adaptation models
Noise measurement
Domain adaptation
deep learning
distribution sampling
Information Security
XU, Xuemiao
HE, Hai
ZHANG, Huaidong
XU, Yangyang
HE, Shengfeng
Unsupervised domain adaptation via importance sampling
description Unsupervised domain adaptation aims to generalize a model from the label-rich source domain to the unlabeled target domain. Existing works mainly focus on aligning the global distribution statistics between source and target domains. However, they neglect distractions from the unexpected noisy samples in domain distribution estimation, leading to domain misalignment or even negative transfer. In this paper, we present an importance sampling method for domain adaptation (ISDA), to measure sample contributions according to their "informative" levels. In particular, informative samples, as well as outliers, can be effectively modeled using feature-norm and prediction entropy of the network. The importance of information is further formulated as the importance sampling losses in features and label spaces. In this way, the proposed model mitigates the noisy outliers while enhancing the important samples during domain alignment. In addition, our model is easy to implement yet effective, and it does not introduce any extra parameters. Extensive experiments on several benchmark datasets show that our method outperforms state-of-the-art methods under both the standard and partial domain adaptation settings.
format text
author XU, Xuemiao
HE, Hai
ZHANG, Huaidong
XU, Yangyang
HE, Shengfeng
author_facet XU, Xuemiao
HE, Hai
ZHANG, Huaidong
XU, Yangyang
HE, Shengfeng
author_sort XU, Xuemiao
title Unsupervised domain adaptation via importance sampling
title_short Unsupervised domain adaptation via importance sampling
title_full Unsupervised domain adaptation via importance sampling
title_fullStr Unsupervised domain adaptation via importance sampling
title_full_unstemmed Unsupervised domain adaptation via importance sampling
title_sort unsupervised domain adaptation via importance sampling
publisher Institutional Knowledge at Singapore Management University
publishDate 2020
url https://ink.library.smu.edu.sg/sis_research/7884
_version_ 1770576575963594752