Revisiting local descriptor for improved few-shot classification

Few-shot classification studies the problem of quickly adapting a deep learner to understanding novel classes based on few support images. In this context, recent research efforts have been aimed at designing more and more complex classifiers that measure similarities between query and support image...

Full description

Saved in:
Bibliographic Details
Main Authors: HE, Jun, HONG, Richang, LIU, Xueliang, XU, Mingliang, SUN, Qianru
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7558
https://ink.library.smu.edu.sg/context/sis_research/article/8561/viewcontent/dcap.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8561
record_format dspace
spelling sg-smu-ink.sis_research-85612022-11-29T07:01:45Z Revisiting local descriptor for improved few-shot classification HE, Jun HONG, Richang LIU, Xueliang XU, Mingliang SUN, Qianru Few-shot classification studies the problem of quickly adapting a deep learner to understanding novel classes based on few support images. In this context, recent research efforts have been aimed at designing more and more complex classifiers that measure similarities between query and support images but left the importance of feature embeddings seldom explored. We show that the reliance on sophisticated classifiers is not necessary, and a simple classifier applied directly to improved feature embeddings can instead outperform most of the leading methods in the literature. To this end, we present a new method, named DCAP, for few-shot classification, in which we investigate how one can improve the quality of embeddings by leveraging Dense Classification and Attentive Pooling (DCAP). Specifically, we propose to train a learner on base classes with abundant samples to solve dense classification problem first and then meta-train the learner on plenty of randomly sampled few-shot tasks to adapt it to few-shot scenario or the test time scenario. During meta-training, we suggest to pool feature maps by applying attentive pooling instead of the widely used global average pooling to prepare embeddings for few-shot classification. Attentive pooling learns to reweight local descriptors, explaining what the learner is looking for as evidence for decision making. Experiments on two benchmark datasets show the proposed method to be superior in multiple few-shot settings while being simpler and more explainable. 2022-06-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7558 info:doi/10.1145/3511917 https://ink.library.smu.edu.sg/context/sis_research/article/8561/viewcontent/dcap.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University few-shot learning image classification visual recognition meta-learning attention networks Databases and Information Systems Graphics and Human Computer Interfaces
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic few-shot learning
image classification
visual recognition
meta-learning
attention networks
Databases and Information Systems
Graphics and Human Computer Interfaces
spellingShingle few-shot learning
image classification
visual recognition
meta-learning
attention networks
Databases and Information Systems
Graphics and Human Computer Interfaces
HE, Jun
HONG, Richang
LIU, Xueliang
XU, Mingliang
SUN, Qianru
Revisiting local descriptor for improved few-shot classification
description Few-shot classification studies the problem of quickly adapting a deep learner to understanding novel classes based on few support images. In this context, recent research efforts have been aimed at designing more and more complex classifiers that measure similarities between query and support images but left the importance of feature embeddings seldom explored. We show that the reliance on sophisticated classifiers is not necessary, and a simple classifier applied directly to improved feature embeddings can instead outperform most of the leading methods in the literature. To this end, we present a new method, named DCAP, for few-shot classification, in which we investigate how one can improve the quality of embeddings by leveraging Dense Classification and Attentive Pooling (DCAP). Specifically, we propose to train a learner on base classes with abundant samples to solve dense classification problem first and then meta-train the learner on plenty of randomly sampled few-shot tasks to adapt it to few-shot scenario or the test time scenario. During meta-training, we suggest to pool feature maps by applying attentive pooling instead of the widely used global average pooling to prepare embeddings for few-shot classification. Attentive pooling learns to reweight local descriptors, explaining what the learner is looking for as evidence for decision making. Experiments on two benchmark datasets show the proposed method to be superior in multiple few-shot settings while being simpler and more explainable.
format text
author HE, Jun
HONG, Richang
LIU, Xueliang
XU, Mingliang
SUN, Qianru
author_facet HE, Jun
HONG, Richang
LIU, Xueliang
XU, Mingliang
SUN, Qianru
author_sort HE, Jun
title Revisiting local descriptor for improved few-shot classification
title_short Revisiting local descriptor for improved few-shot classification
title_full Revisiting local descriptor for improved few-shot classification
title_fullStr Revisiting local descriptor for improved few-shot classification
title_full_unstemmed Revisiting local descriptor for improved few-shot classification
title_sort revisiting local descriptor for improved few-shot classification
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/7558
https://ink.library.smu.edu.sg/context/sis_research/article/8561/viewcontent/dcap.pdf
_version_ 1770576372361592832