Disentangled feature representation for few-shot image classification
Learning the generalizable feature representation is critical to few-shot image classification. While recent works exploited task-specific feature embedding using meta-tasks for few-shot learning, they are limited in many challenging tasks as being distracted by the excursive features such as the ba...
Saved in:
Main Authors: | , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/170577 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-170577 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1705772023-09-19T08:32:14Z Disentangled feature representation for few-shot image classification Cheng, Hao Wang, Yufei Li, Haoliang Kot, Alex Chichung Wen, Bihan School of Electrical and Electronic Engineering Engineering::Electrical and electronic engineering Task Analysis Feature Extraction Learning the generalizable feature representation is critical to few-shot image classification. While recent works exploited task-specific feature embedding using meta-tasks for few-shot learning, they are limited in many challenging tasks as being distracted by the excursive features such as the background, domain, and style of the image samples. In this work, we propose a novel disentangled feature representation (DFR) framework, dubbed DFR, for few-shot learning applications. DFR can adaptively decouple the discriminative features that are modeled by the classification branch, from the class-irrelevant component of the variation branch. In general, most of the popular deep few-shot learning methods can be plugged in as the classification branch, thus DFR can boost their performance on various few-shot tasks. Furthermore, we propose a novel FS-DomainNet dataset based on DomainNet, for benchmarking the few-shot domain generalization (DG) tasks. We conducted extensive experiments to evaluate the proposed DFR on general, fine-grained, and cross-domain few-shot classification, as well as few-shot DG, using the corresponding four benchmarks, i.e., mini-ImageNet, tiered-ImageNet, Caltech-UCSD Birds 200-2011 (CUB), and the proposed FS-DomainNet. Thanks to the effective feature disentangling, the DFR-based few-shot classifiers achieved state-of-the-art results on all datasets. Ministry of Education (MOE) This work was supported in part by the Singapore Ministry of Education Academic Research Fund Tier 1 under Grant RG61/22 and Startup Grant, in part by the Research Grant Council (RGC) of Hong Kong through Early Career Scheme (ECS) under Grant 21200522, and in part by the Sichuan Science and Technology Program under Grant 2022NSFSC0551. 2023-09-19T08:32:14Z 2023-09-19T08:32:14Z 2023 Journal Article Cheng, H., Wang, Y., Li, H., Kot, A. C. & Wen, B. (2023). Disentangled feature representation for few-shot image classification. IEEE Transactions On Neural Networks and Learning Systems. https://dx.doi.org/10.1109/TNNLS.2023.3241919 2162-237X https://hdl.handle.net/10356/170577 10.1109/TNNLS.2023.3241919 37027772 2-s2.0-85149374301 en RG61/22 IEEE Transactions on Neural Networks and Learning Systems © 2023 IEEE. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Electrical and electronic engineering Task Analysis Feature Extraction |
spellingShingle |
Engineering::Electrical and electronic engineering Task Analysis Feature Extraction Cheng, Hao Wang, Yufei Li, Haoliang Kot, Alex Chichung Wen, Bihan Disentangled feature representation for few-shot image classification |
description |
Learning the generalizable feature representation is critical to few-shot image classification. While recent works exploited task-specific feature embedding using meta-tasks for few-shot learning, they are limited in many challenging tasks as being distracted by the excursive features such as the background, domain, and style of the image samples. In this work, we propose a novel disentangled feature representation (DFR) framework, dubbed DFR, for few-shot learning applications. DFR can adaptively decouple the discriminative features that are modeled by the classification branch, from the class-irrelevant component of the variation branch. In general, most of the popular deep few-shot learning methods can be plugged in as the classification branch, thus DFR can boost their performance on various few-shot tasks. Furthermore, we propose a novel FS-DomainNet dataset based on DomainNet, for benchmarking the few-shot domain generalization (DG) tasks. We conducted extensive experiments to evaluate the proposed DFR on general, fine-grained, and cross-domain few-shot classification, as well as few-shot DG, using the corresponding four benchmarks, i.e., mini-ImageNet, tiered-ImageNet, Caltech-UCSD Birds 200-2011 (CUB), and the proposed FS-DomainNet. Thanks to the effective feature disentangling, the DFR-based few-shot classifiers achieved state-of-the-art results on all datasets. |
author2 |
School of Electrical and Electronic Engineering |
author_facet |
School of Electrical and Electronic Engineering Cheng, Hao Wang, Yufei Li, Haoliang Kot, Alex Chichung Wen, Bihan |
format |
Article |
author |
Cheng, Hao Wang, Yufei Li, Haoliang Kot, Alex Chichung Wen, Bihan |
author_sort |
Cheng, Hao |
title |
Disentangled feature representation for few-shot image classification |
title_short |
Disentangled feature representation for few-shot image classification |
title_full |
Disentangled feature representation for few-shot image classification |
title_fullStr |
Disentangled feature representation for few-shot image classification |
title_full_unstemmed |
Disentangled feature representation for few-shot image classification |
title_sort |
disentangled feature representation for few-shot image classification |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/170577 |
_version_ |
1779156305707008000 |