Transductive transfer learning for visual recognition
In recent years, deep neural networks (DNNs) have brought great advances to various computer vision tasks, such as image classification, object detection, semantic segmentation, etc. However, the considerable successes of DNNs are achieved at a high cost of quite a lot of densely labeled training im...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Doctor of Philosophy |
Language: | English |
Published: |
Nanyang Technological University
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/164573 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
Summary: | In recent years, deep neural networks (DNNs) have brought great advances to various computer vision tasks, such as image classification, object detection, semantic segmentation, etc. However, the considerable successes of DNNs are achieved at a high cost of quite a lot of densely labeled training images that are extremely costly and laborious to establish. One way of circumventing such a limitation is to utilize the annotated images from existing related datasets (called the “source domain”) in network training. Unfortunately, the DNNs trained on source domains often undergo a drastic performance degradation when applied to the “target domain” due to the distribution mismatch. In such scenarios, transfer learning (or called knowledge transfer) between domains is desirable and necessary. In this thesis, we explore Transductive Transfer Learning for visual recognition, where the data distributions of labeled source and unlabeled target domains are different while the source and target tasks are the same. More specifically, we investigate three representative types of transductive transfer learning, including domain generalization, unsupervised domain adaptation and source-free unsupervised domain adaptation. In domain generalization, given labeled source domain data, the goal is to learn a generalized visual recognition model that well performs over unseen target-domain data. In other words, domain generalization aims to learn domain invariant features (or transferable features) without requiring target-domain data in training. In this thesis, we proposed a novel domain generalization approach that effectively randomizes source-domain images in frequency space, which encourages DNNs to learn style-invariant visual features that generalize well to unseen target domains. In unsupervised domain adaptation, given labeled source-domain data and unla beled target-domain data, the goal is to learn an adaptive visual recognition model that well performs over target-domain data. Different from domain generalization, in the transfer learning setup of unsupervised domain adaptation, the unlabeled target-domain data is accessible during training. Therefore, unsupervised domain adaptation largely focuses on exploiting unlabelled target-domain data to improve network performance. In this thesis, we developed four novel unsupervised domain adaptation techniques that effectively transfer knowledge from labeled source domains to the unlabeled target domain. More specifically, we designed different unsupervised losses on unlabeled target-domain data for learning a well-performed model in the target domain. In source-free unsupervised domain adaptation, given a source-trained model and unlabeled target-domain data, the goal is to adapt the source-trained model to per form well on unlabeled target-domain data. Different from unsupervised domain adaptation, in the transfer learning setup of source-free unsupervised domain adaptation, the labeled source-domain data is not accessible during training, where we aim to adapt source-trained models to fit target data distribution without accessing the source-domain data. Under a such transfer learning setup, the only information carried forward is a portable source-trained model, which largely alleviates the concern of data privacy, data portability and data transmission efficiency. To this end, we proposed a novel source-free unsupervised domain adaptation approach that exploits historical source hypothesis to make up for the absence of source-domain data in this transfer learning setup. Experimental results over various visual recognition benchmarks indicate our pro posed transfer learning approaches achieve superior performance, enabling transferring DNNs across different domains. |
---|