How important is the train-validation split in meta-learning?

Meta-learning aims to perform fast adaptation on a new task through learning a “prior” from multiple existing tasks. A common practice in meta-learning is to perform a train-validation split (train-val method) where the prior adapts to the task on one split of the data, and the resulting predictor i...

Full description

Saved in:
Bibliographic Details
Main Authors: BAI, Yu, CHEN, Minshuo, ZHOU, Pan, ZHAO, Tuo, LEE, D. Jason, KAKADE, Sham, WANG, Huan, XIONG, Caiming
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2021
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8991
https://ink.library.smu.edu.sg/context/sis_research/article/9994/viewcontent/2021_ICML_Metalearning.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9994
record_format dspace
spelling sg-smu-ink.sis_research-99942024-07-25T08:26:25Z How important is the train-validation split in meta-learning? BAI, Yu CHEN, Minshuo ZHOU, Pan ZHAO, Tuo LEE, D. Jason KAKADE, Sham WANG, Huan XIONG, Caiming Meta-learning aims to perform fast adaptation on a new task through learning a “prior” from multiple existing tasks. A common practice in meta-learning is to perform a train-validation split (train-val method) where the prior adapts to the task on one split of the data, and the resulting predictor is evaluated on another split. Despite its prevalence, the importance of the train-validation split is not well understood either in theory or in practice, particularly in comparison to the more direct train-train method, which uses all the pertask data for both training and evaluation. We provide a detailed theoretical study on whether and when the train-validation split is helpful in the linear centroid meta-learning problem. In the agnostic case, we show that the expected loss of the train-val method is minimized at the optimal prior for meta testing, and this is not the case for the train-train method in general without structural assumptions on the data. In contrast, in the realizable case where the data are generated from linear models, we show that both the train-val and train-train losses are minimized at the optimal prior in expectation. Further, perhaps surprisingly, our main result shows that the train-train method achieves a strictly better excess loss in this realizable case, even when the regularization parameter and split ratio are optimally tuned for both methods. Our results highlight that sample splitting may not always be preferable, especially when the data is realizable by the model. We validate our theories by experimentally showing that the train-train method can indeed outperform the train-val method, on both simulations and real meta-learning tasks. 2021-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8991 https://ink.library.smu.edu.sg/context/sis_research/article/9994/viewcontent/2021_ICML_Metalearning.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Artificial Intelligence and Robotics Graphics and Human Computer Interfaces
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Artificial Intelligence and Robotics
Graphics and Human Computer Interfaces
spellingShingle Artificial Intelligence and Robotics
Graphics and Human Computer Interfaces
BAI, Yu
CHEN, Minshuo
ZHOU, Pan
ZHAO, Tuo
LEE, D. Jason
KAKADE, Sham
WANG, Huan
XIONG, Caiming
How important is the train-validation split in meta-learning?
description Meta-learning aims to perform fast adaptation on a new task through learning a “prior” from multiple existing tasks. A common practice in meta-learning is to perform a train-validation split (train-val method) where the prior adapts to the task on one split of the data, and the resulting predictor is evaluated on another split. Despite its prevalence, the importance of the train-validation split is not well understood either in theory or in practice, particularly in comparison to the more direct train-train method, which uses all the pertask data for both training and evaluation. We provide a detailed theoretical study on whether and when the train-validation split is helpful in the linear centroid meta-learning problem. In the agnostic case, we show that the expected loss of the train-val method is minimized at the optimal prior for meta testing, and this is not the case for the train-train method in general without structural assumptions on the data. In contrast, in the realizable case where the data are generated from linear models, we show that both the train-val and train-train losses are minimized at the optimal prior in expectation. Further, perhaps surprisingly, our main result shows that the train-train method achieves a strictly better excess loss in this realizable case, even when the regularization parameter and split ratio are optimally tuned for both methods. Our results highlight that sample splitting may not always be preferable, especially when the data is realizable by the model. We validate our theories by experimentally showing that the train-train method can indeed outperform the train-val method, on both simulations and real meta-learning tasks.
format text
author BAI, Yu
CHEN, Minshuo
ZHOU, Pan
ZHAO, Tuo
LEE, D. Jason
KAKADE, Sham
WANG, Huan
XIONG, Caiming
author_facet BAI, Yu
CHEN, Minshuo
ZHOU, Pan
ZHAO, Tuo
LEE, D. Jason
KAKADE, Sham
WANG, Huan
XIONG, Caiming
author_sort BAI, Yu
title How important is the train-validation split in meta-learning?
title_short How important is the train-validation split in meta-learning?
title_full How important is the train-validation split in meta-learning?
title_fullStr How important is the train-validation split in meta-learning?
title_full_unstemmed How important is the train-validation split in meta-learning?
title_sort how important is the train-validation split in meta-learning?
publisher Institutional Knowledge at Singapore Management University
publishDate 2021
url https://ink.library.smu.edu.sg/sis_research/8991
https://ink.library.smu.edu.sg/context/sis_research/article/9994/viewcontent/2021_ICML_Metalearning.pdf
_version_ 1814047702523052032