Improving feature generalizability with multitask learning in class incremental learning

Many deep learning applications, like keyword spotting [1], [2], require the incorporation of new concepts (classes) over time, referred to as Class Incremental Learning (CIL). The major challenge in CIL is catastrophic forgetting, i.e., preserving as much of the old knowledge as possible while lear...

Full description

Saved in:
Bibliographic Details
Main Authors: MA, Dong, TANG, Chi Ian, MASCOLO, Cecilia
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7305
https://ink.library.smu.edu.sg/context/sis_research/article/8308/viewcontent/2204.12915.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8308
record_format dspace
spelling sg-smu-ink.sis_research-83082022-09-29T07:35:54Z Improving feature generalizability with multitask learning in class incremental learning MA, Dong TANG, Chi Ian MASCOLO, Cecilia Many deep learning applications, like keyword spotting [1], [2], require the incorporation of new concepts (classes) over time, referred to as Class Incremental Learning (CIL). The major challenge in CIL is catastrophic forgetting, i.e., preserving as much of the old knowledge as possible while learning new tasks. Various techniques, such as regularization, knowledge distillation, and the use of exemplars, have been proposed to resolve this issue. However, prior works primarily focus on the incremental learning step, while ignoring the optimization during the base model training. We hypothesise that a more transferable and generalizable feature representation from the base model would be beneficial to incremental learning.In this work, we adopt multitask learning during base model training to improve the feature generalizability. Specifically, instead of training a single model with all the base classes, we decompose the base classes into multiple subsets and regard each of them as a task. These tasks are trained concurrently and a shared feature extractor is obtained for incremental learning. We evaluate our approach on two datasets under various configurations. The results show that our approach enhances the average incremental learning accuracy by up to 5.5%, which enables more reliable and accurate keyword spotting over time. Moreover, the proposed approach can be combined with many existing techniques and provides additional performance gain. 2022-04-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7305 info:doi/10.1109/ICASSP43922.2022.9746862 https://ink.library.smu.edu.sg/context/sis_research/article/8308/viewcontent/2204.12915.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Class Incremental Learning Continual Learning Multitask Learning Keyword Spotting Artificial Intelligence and Robotics Computer Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Class Incremental Learning
Continual Learning
Multitask Learning
Keyword Spotting
Artificial Intelligence and Robotics
Computer Engineering
spellingShingle Class Incremental Learning
Continual Learning
Multitask Learning
Keyword Spotting
Artificial Intelligence and Robotics
Computer Engineering
MA, Dong
TANG, Chi Ian
MASCOLO, Cecilia
Improving feature generalizability with multitask learning in class incremental learning
description Many deep learning applications, like keyword spotting [1], [2], require the incorporation of new concepts (classes) over time, referred to as Class Incremental Learning (CIL). The major challenge in CIL is catastrophic forgetting, i.e., preserving as much of the old knowledge as possible while learning new tasks. Various techniques, such as regularization, knowledge distillation, and the use of exemplars, have been proposed to resolve this issue. However, prior works primarily focus on the incremental learning step, while ignoring the optimization during the base model training. We hypothesise that a more transferable and generalizable feature representation from the base model would be beneficial to incremental learning.In this work, we adopt multitask learning during base model training to improve the feature generalizability. Specifically, instead of training a single model with all the base classes, we decompose the base classes into multiple subsets and regard each of them as a task. These tasks are trained concurrently and a shared feature extractor is obtained for incremental learning. We evaluate our approach on two datasets under various configurations. The results show that our approach enhances the average incremental learning accuracy by up to 5.5%, which enables more reliable and accurate keyword spotting over time. Moreover, the proposed approach can be combined with many existing techniques and provides additional performance gain.
format text
author MA, Dong
TANG, Chi Ian
MASCOLO, Cecilia
author_facet MA, Dong
TANG, Chi Ian
MASCOLO, Cecilia
author_sort MA, Dong
title Improving feature generalizability with multitask learning in class incremental learning
title_short Improving feature generalizability with multitask learning in class incremental learning
title_full Improving feature generalizability with multitask learning in class incremental learning
title_fullStr Improving feature generalizability with multitask learning in class incremental learning
title_full_unstemmed Improving feature generalizability with multitask learning in class incremental learning
title_sort improving feature generalizability with multitask learning in class incremental learning
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/7305
https://ink.library.smu.edu.sg/context/sis_research/article/8308/viewcontent/2204.12915.pdf
_version_ 1770576308231733248