Clustered task-aware meta-learning by learning from learning paths
To enable effective learning of new tasks with only a few examples, meta-learning acquires common knowledge from the existing tasks with a globally shared meta-learner. To further address the problem of task heterogeneity, recent developments balance between customization and generalization by incor...
Saved in:
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/172181 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-172181 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1721812023-11-28T06:12:48Z Clustered task-aware meta-learning by learning from learning paths Peng, Danni Pan, Sinno Jialin School of Computer Science and Engineering Engineering::Computer science and engineering Task Clustering Task-Aware Meta-Learning To enable effective learning of new tasks with only a few examples, meta-learning acquires common knowledge from the existing tasks with a globally shared meta-learner. To further address the problem of task heterogeneity, recent developments balance between customization and generalization by incorporating task clustering to generate task-aware modulation to be applied to the global meta-learner. However, these methods learn task representation mostly from the features ofinput data, while the task-specific optimization process with respect to the base-learner is often neglected. In this work, we propose a Clustered Task-Aware Meta-Learning (CTML) framework with task representation learned from both features and learning paths. We first conduct rehearsed task learning from the common initialization, and collect a set of geometric quantities that adequately describes this learning path. By inputting this set of values into a meta path learner, we automatically abstract path representation optimized for downstream clustering and modulation. Aggregating the path and feature representations results in an improved task representation. To further improve inference efficiency, we devise a shortcut tunnel to bypass the rehearsed learning process at a meta-testing time. Extensive experiments on two real-world application domains: few-shot image classification and cold-start recommendation demonstrate the superiority of CTML compared to state-of-the-art methods. We provide our code at https://github.com/didiya0825. This work was supported in part by Alibaba Group through Alibaba Innovative Research (AIR) Program and Alibaba-NTU Singapore Joint Research Institute (JRI), Nanyang Technological University, Singapore. 2023-11-28T06:12:48Z 2023-11-28T06:12:48Z 2023 Journal Article Peng, D. & Pan, S. J. (2023). Clustered task-aware meta-learning by learning from learning paths. IEEE Transactions On Pattern Analysis and Machine Intelligence, 45(8), 9426-9438. https://dx.doi.org/10.1109/TPAMI.2023.3250323 0162-8828 https://hdl.handle.net/10356/172181 10.1109/TPAMI.2023.3250323 37028045 2-s2.0-85149836176 8 45 9426 9438 en IEEE transactions on pattern analysis and machine intelligence © 2023 IEEE. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Task Clustering Task-Aware Meta-Learning |
spellingShingle |
Engineering::Computer science and engineering Task Clustering Task-Aware Meta-Learning Peng, Danni Pan, Sinno Jialin Clustered task-aware meta-learning by learning from learning paths |
description |
To enable effective learning of new tasks with only a few examples, meta-learning acquires common knowledge from the existing tasks with a globally shared meta-learner. To further address the problem of task heterogeneity, recent developments balance between customization and generalization by incorporating task clustering to generate task-aware modulation to be applied to the global meta-learner. However, these methods learn task representation mostly from the features ofinput data, while the task-specific optimization process with respect to the base-learner is often neglected. In this work, we propose a Clustered Task-Aware Meta-Learning (CTML) framework with task representation learned from both features and learning paths. We first conduct rehearsed task learning from the common initialization, and collect a set of geometric quantities that adequately describes this learning path. By inputting this set of values into a meta path learner, we automatically abstract path representation optimized for downstream clustering and modulation. Aggregating the path and feature representations results in an improved task representation. To further improve inference efficiency, we devise a shortcut tunnel to bypass the rehearsed learning process at a meta-testing time. Extensive experiments on two real-world application domains: few-shot image classification and cold-start recommendation demonstrate the superiority of CTML compared to state-of-the-art methods. We provide our code at https://github.com/didiya0825. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Peng, Danni Pan, Sinno Jialin |
format |
Article |
author |
Peng, Danni Pan, Sinno Jialin |
author_sort |
Peng, Danni |
title |
Clustered task-aware meta-learning by learning from learning paths |
title_short |
Clustered task-aware meta-learning by learning from learning paths |
title_full |
Clustered task-aware meta-learning by learning from learning paths |
title_fullStr |
Clustered task-aware meta-learning by learning from learning paths |
title_full_unstemmed |
Clustered task-aware meta-learning by learning from learning paths |
title_sort |
clustered task-aware meta-learning by learning from learning paths |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/172181 |
_version_ |
1783955627964366848 |