Unsupervised domain adaptation in the wild via disentangling representation learning

Most recently proposed unsupervised domain adaptation algorithms attempt to learn domain invariant features by confusing a domain classifier through adversarial training. In this paper, we argue that this may not be an optimal solution in the real-world setting (a.k.a. in the wild) as the difference...

Full description

Saved in:
Bibliographic Details
Main Authors: Li, Haoliang, Wan, Renjie, Wang, Shiqi, Kot, Alex Chichung
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2022
Subjects:
Online Access:https://hdl.handle.net/10356/160950
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-160950
record_format dspace
spelling sg-ntu-dr.10356-1609502022-08-08T07:33:34Z Unsupervised domain adaptation in the wild via disentangling representation learning Li, Haoliang Wan, Renjie Wang, Shiqi Kot, Alex Chichung School of Electrical and Electronic Engineering Interdisciplinary Graduate School (IGS) Rapid-Rich Object Search (ROSE) Lab Engineering::Electrical and electronic engineering In the Wild Cross-Domain Most recently proposed unsupervised domain adaptation algorithms attempt to learn domain invariant features by confusing a domain classifier through adversarial training. In this paper, we argue that this may not be an optimal solution in the real-world setting (a.k.a. in the wild) as the difference in terms of label information between domains has been largely ignored. As labeled instances are not available in the target domain in unsupervised domain adaptation tasks, it is difficult to explicitly capture the label difference between domains. To address this issue, we propose to learn a disentangled latent representation based on implicit autoencoders. In particular, a latent representation is disentangled into a global code and a local code. The global code is capturing category information via an encoder with a prior, and the local code is transferable across domains, which captures the “style” related information via an implicit decoder. Experimental results on digit recognition, object recognition and semantic segmentation demonstrate the effectiveness of our proposed method. Nanyang Technological University This research is supported in part by the Wallenberg-NTU Presidential Postdoctoral Fellowship, the NTU-PKU Joint Research Institute, a collaboration between the Nanyang Technological University and Peking University that is sponsored by a donation from the Ng Teng Fong Charitable Foundation, and the Science and Technology Foundation of Guangzhou Huangpu Development District under Grant 201902010028. 2022-08-08T07:15:42Z 2022-08-08T07:15:42Z 2021 Journal Article Li, H., Wan, R., Wang, S. & Kot, A. C. (2021). Unsupervised domain adaptation in the wild via disentangling representation learning. International Journal of Computer Vision, 129(2), 267-283. https://dx.doi.org/10.1007/s11263-020-01364-5 0920-5691 https://hdl.handle.net/10356/160950 10.1007/s11263-020-01364-5 2-s2.0-85089294127 2 129 267 283 en International Journal of Computer Vision © 2020 Springer Science+Business Media, LLC, part of Springer Nature. All rights reserved.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
In the Wild
Cross-Domain
spellingShingle Engineering::Electrical and electronic engineering
In the Wild
Cross-Domain
Li, Haoliang
Wan, Renjie
Wang, Shiqi
Kot, Alex Chichung
Unsupervised domain adaptation in the wild via disentangling representation learning
description Most recently proposed unsupervised domain adaptation algorithms attempt to learn domain invariant features by confusing a domain classifier through adversarial training. In this paper, we argue that this may not be an optimal solution in the real-world setting (a.k.a. in the wild) as the difference in terms of label information between domains has been largely ignored. As labeled instances are not available in the target domain in unsupervised domain adaptation tasks, it is difficult to explicitly capture the label difference between domains. To address this issue, we propose to learn a disentangled latent representation based on implicit autoencoders. In particular, a latent representation is disentangled into a global code and a local code. The global code is capturing category information via an encoder with a prior, and the local code is transferable across domains, which captures the “style” related information via an implicit decoder. Experimental results on digit recognition, object recognition and semantic segmentation demonstrate the effectiveness of our proposed method.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Li, Haoliang
Wan, Renjie
Wang, Shiqi
Kot, Alex Chichung
format Article
author Li, Haoliang
Wan, Renjie
Wang, Shiqi
Kot, Alex Chichung
author_sort Li, Haoliang
title Unsupervised domain adaptation in the wild via disentangling representation learning
title_short Unsupervised domain adaptation in the wild via disentangling representation learning
title_full Unsupervised domain adaptation in the wild via disentangling representation learning
title_fullStr Unsupervised domain adaptation in the wild via disentangling representation learning
title_full_unstemmed Unsupervised domain adaptation in the wild via disentangling representation learning
title_sort unsupervised domain adaptation in the wild via disentangling representation learning
publishDate 2022
url https://hdl.handle.net/10356/160950
_version_ 1743119591088848896