Disentangled feature representation for few-shot image classification

Learning the generalizable feature representation is critical to few-shot image classification. While recent works exploited task-specific feature embedding using meta-tasks for few-shot learning, they are limited in many challenging tasks as being distracted by the excursive features such as the ba...

Full description

Saved in:
Bibliographic Details
Main Authors: Cheng, Hao, Wang, Yufei, Li, Haoliang, Kot, Alex Chichung, Wen, Bihan
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2023
Subjects:
Online Access:https://hdl.handle.net/10356/170577
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-170577
record_format dspace
spelling sg-ntu-dr.10356-1705772023-09-19T08:32:14Z Disentangled feature representation for few-shot image classification Cheng, Hao Wang, Yufei Li, Haoliang Kot, Alex Chichung Wen, Bihan School of Electrical and Electronic Engineering Engineering::Electrical and electronic engineering Task Analysis Feature Extraction Learning the generalizable feature representation is critical to few-shot image classification. While recent works exploited task-specific feature embedding using meta-tasks for few-shot learning, they are limited in many challenging tasks as being distracted by the excursive features such as the background, domain, and style of the image samples. In this work, we propose a novel disentangled feature representation (DFR) framework, dubbed DFR, for few-shot learning applications. DFR can adaptively decouple the discriminative features that are modeled by the classification branch, from the class-irrelevant component of the variation branch. In general, most of the popular deep few-shot learning methods can be plugged in as the classification branch, thus DFR can boost their performance on various few-shot tasks. Furthermore, we propose a novel FS-DomainNet dataset based on DomainNet, for benchmarking the few-shot domain generalization (DG) tasks. We conducted extensive experiments to evaluate the proposed DFR on general, fine-grained, and cross-domain few-shot classification, as well as few-shot DG, using the corresponding four benchmarks, i.e., mini-ImageNet, tiered-ImageNet, Caltech-UCSD Birds 200-2011 (CUB), and the proposed FS-DomainNet. Thanks to the effective feature disentangling, the DFR-based few-shot classifiers achieved state-of-the-art results on all datasets. Ministry of Education (MOE) This work was supported in part by the Singapore Ministry of Education Academic Research Fund Tier 1 under Grant RG61/22 and Startup Grant, in part by the Research Grant Council (RGC) of Hong Kong through Early Career Scheme (ECS) under Grant 21200522, and in part by the Sichuan Science and Technology Program under Grant 2022NSFSC0551. 2023-09-19T08:32:14Z 2023-09-19T08:32:14Z 2023 Journal Article Cheng, H., Wang, Y., Li, H., Kot, A. C. & Wen, B. (2023). Disentangled feature representation for few-shot image classification. IEEE Transactions On Neural Networks and Learning Systems. https://dx.doi.org/10.1109/TNNLS.2023.3241919 2162-237X https://hdl.handle.net/10356/170577 10.1109/TNNLS.2023.3241919 37027772 2-s2.0-85149374301 en RG61/22 IEEE Transactions on Neural Networks and Learning Systems © 2023 IEEE. All rights reserved.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
Task Analysis
Feature Extraction
spellingShingle Engineering::Electrical and electronic engineering
Task Analysis
Feature Extraction
Cheng, Hao
Wang, Yufei
Li, Haoliang
Kot, Alex Chichung
Wen, Bihan
Disentangled feature representation for few-shot image classification
description Learning the generalizable feature representation is critical to few-shot image classification. While recent works exploited task-specific feature embedding using meta-tasks for few-shot learning, they are limited in many challenging tasks as being distracted by the excursive features such as the background, domain, and style of the image samples. In this work, we propose a novel disentangled feature representation (DFR) framework, dubbed DFR, for few-shot learning applications. DFR can adaptively decouple the discriminative features that are modeled by the classification branch, from the class-irrelevant component of the variation branch. In general, most of the popular deep few-shot learning methods can be plugged in as the classification branch, thus DFR can boost their performance on various few-shot tasks. Furthermore, we propose a novel FS-DomainNet dataset based on DomainNet, for benchmarking the few-shot domain generalization (DG) tasks. We conducted extensive experiments to evaluate the proposed DFR on general, fine-grained, and cross-domain few-shot classification, as well as few-shot DG, using the corresponding four benchmarks, i.e., mini-ImageNet, tiered-ImageNet, Caltech-UCSD Birds 200-2011 (CUB), and the proposed FS-DomainNet. Thanks to the effective feature disentangling, the DFR-based few-shot classifiers achieved state-of-the-art results on all datasets.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Cheng, Hao
Wang, Yufei
Li, Haoliang
Kot, Alex Chichung
Wen, Bihan
format Article
author Cheng, Hao
Wang, Yufei
Li, Haoliang
Kot, Alex Chichung
Wen, Bihan
author_sort Cheng, Hao
title Disentangled feature representation for few-shot image classification
title_short Disentangled feature representation for few-shot image classification
title_full Disentangled feature representation for few-shot image classification
title_fullStr Disentangled feature representation for few-shot image classification
title_full_unstemmed Disentangled feature representation for few-shot image classification
title_sort disentangled feature representation for few-shot image classification
publishDate 2023
url https://hdl.handle.net/10356/170577
_version_ 1779156305707008000