Efficient meta learning via minibatch proximal update

We address the problem of meta-learning which learns a prior over hypothesis from a sample of meta-training tasks for fast adaptation on meta-testing tasks. A particularly simple yet successful paradigm for this research is model-agnostic meta-learning (MAML). Implementation and analysis of MAML, ho...

Full description

Saved in:
Bibliographic Details
Main Authors: ZHOU, Pan, YUAN, Xiao-Tong, XU, Huan, YAN, Shuicheng, FENG, Jiashi
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2019
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8995
https://ink.library.smu.edu.sg/context/sis_research/article/9998/viewcontent/2019_NeurIPS_efficient_meta_learning.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9998
record_format dspace
spelling sg-smu-ink.sis_research-99982024-07-25T08:24:41Z Efficient meta learning via minibatch proximal update ZHOU, Pan YUAN, Xiao-Tong XU, Huan YAN, Shuicheng FENG, Jiashi We address the problem of meta-learning which learns a prior over hypothesis from a sample of meta-training tasks for fast adaptation on meta-testing tasks. A particularly simple yet successful paradigm for this research is model-agnostic meta-learning (MAML). Implementation and analysis of MAML, however, can be tricky; first-order approximation is usually adopted to avoid directly computing Hessian matrix but as a result the convergence and generalization guarantees remain largely mysterious for MAML. To remedy this deficiency, in this paper we propose a minibatch proximal update based meta-learning approach for learning to efficient hypothesis transfer. The principle is to learn a prior hypothesis shared across tasks such that the minibatch risk minimization biased regularized by this prior can quickly converge to the optimal hypothesis in each training task. The prior hypothesis training model can be efficiently optimized via SGD with provable convergence guarantees for both convex and non-convex problems. Moreover, we theoretically justify the benefit of the learnt prior hypothesis for fast adaptation to new few-shot learning tasks via minibatch proximal update. Experimental results on several few-shot regression and classification tasks demonstrate the advantages of our method over state-of-the-arts. 2019-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8995 https://ink.library.smu.edu.sg/context/sis_research/article/9998/viewcontent/2019_NeurIPS_efficient_meta_learning.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Graphics and Human Computer Interfaces
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Graphics and Human Computer Interfaces
spellingShingle Graphics and Human Computer Interfaces
ZHOU, Pan
YUAN, Xiao-Tong
XU, Huan
YAN, Shuicheng
FENG, Jiashi
Efficient meta learning via minibatch proximal update
description We address the problem of meta-learning which learns a prior over hypothesis from a sample of meta-training tasks for fast adaptation on meta-testing tasks. A particularly simple yet successful paradigm for this research is model-agnostic meta-learning (MAML). Implementation and analysis of MAML, however, can be tricky; first-order approximation is usually adopted to avoid directly computing Hessian matrix but as a result the convergence and generalization guarantees remain largely mysterious for MAML. To remedy this deficiency, in this paper we propose a minibatch proximal update based meta-learning approach for learning to efficient hypothesis transfer. The principle is to learn a prior hypothesis shared across tasks such that the minibatch risk minimization biased regularized by this prior can quickly converge to the optimal hypothesis in each training task. The prior hypothesis training model can be efficiently optimized via SGD with provable convergence guarantees for both convex and non-convex problems. Moreover, we theoretically justify the benefit of the learnt prior hypothesis for fast adaptation to new few-shot learning tasks via minibatch proximal update. Experimental results on several few-shot regression and classification tasks demonstrate the advantages of our method over state-of-the-arts.
format text
author ZHOU, Pan
YUAN, Xiao-Tong
XU, Huan
YAN, Shuicheng
FENG, Jiashi
author_facet ZHOU, Pan
YUAN, Xiao-Tong
XU, Huan
YAN, Shuicheng
FENG, Jiashi
author_sort ZHOU, Pan
title Efficient meta learning via minibatch proximal update
title_short Efficient meta learning via minibatch proximal update
title_full Efficient meta learning via minibatch proximal update
title_fullStr Efficient meta learning via minibatch proximal update
title_full_unstemmed Efficient meta learning via minibatch proximal update
title_sort efficient meta learning via minibatch proximal update
publisher Institutional Knowledge at Singapore Management University
publishDate 2019
url https://ink.library.smu.edu.sg/sis_research/8995
https://ink.library.smu.edu.sg/context/sis_research/article/9998/viewcontent/2019_NeurIPS_efficient_meta_learning.pdf
_version_ 1814047703623008256