Fine-grained generalization analysis of vector-valued learning
Many fundamental machine learning tasks can be formulated as a problem of learning with vector-valued functions, where we learn multiple scalar-valued functions together. Although there is some generalization analysis on different specific algorithms under the empirical risk minimization principle,...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2021
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/7203 https://ink.library.smu.edu.sg/context/sis_research/article/8206/viewcontent/vector_val.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-8206 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-82062022-08-04T08:49:28Z Fine-grained generalization analysis of vector-valued learning WU, Liang LEDENT, Antoine LEI, Yunwen KLOFT, Marius Many fundamental machine learning tasks can be formulated as a problem of learning with vector-valued functions, where we learn multiple scalar-valued functions together. Although there is some generalization analysis on different specific algorithms under the empirical risk minimization principle, a unifying analysis of vector-valued learning under a regularization framework is still lacking. In this paper, we initiate the generalization analysis of regularized vector-valued learning algorithms by presenting bounds with a mild dependency on the output dimension and a fast rate on the sample size. Our discussions relax the existing assumptions on the restrictive constraint of hypothesis spaces, smoothness of loss functions and low-noise condition. To understand the interaction between optimization and learning, we further use our results to derive the first generalization bounds for stochastic gradient descent with vector-valued functions. We apply our general results to multi-class classification and multi-label classification, which yield the first bounds with a logarithmic dependency on the output dimension for extreme multi-label classification with the Frobenius regularization. As a byproduct, we derive a Rademacher complexity bound for loss function classes defined in terms of a general strongly convex function. 2021-02-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7203 https://ink.library.smu.edu.sg/context/sis_research/article/8206/viewcontent/vector_val.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Statistical Learning Theory Multi-label Learning Stochastic Gradient Descent Artificial Intelligence and Robotics Theory and Algorithms |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Statistical Learning Theory Multi-label Learning Stochastic Gradient Descent Artificial Intelligence and Robotics Theory and Algorithms |
spellingShingle |
Statistical Learning Theory Multi-label Learning Stochastic Gradient Descent Artificial Intelligence and Robotics Theory and Algorithms WU, Liang LEDENT, Antoine LEI, Yunwen KLOFT, Marius Fine-grained generalization analysis of vector-valued learning |
description |
Many fundamental machine learning tasks can be formulated as a problem of learning with vector-valued functions, where we learn multiple scalar-valued functions together. Although there is some generalization analysis on different specific algorithms under the empirical risk minimization principle, a unifying analysis of vector-valued learning under a regularization framework is still lacking. In this paper, we initiate the generalization analysis of regularized vector-valued learning algorithms by presenting bounds with a mild dependency on the output dimension and a fast rate on the sample size. Our discussions relax the existing assumptions on the restrictive constraint of hypothesis spaces, smoothness of loss functions and low-noise condition. To understand the interaction between optimization and learning, we further use our results to derive the first generalization bounds for stochastic gradient descent with vector-valued functions. We apply our general results to multi-class classification and multi-label classification, which yield the first bounds with a logarithmic dependency on the output dimension for extreme multi-label classification with the Frobenius regularization. As a byproduct, we derive a Rademacher complexity bound for loss function classes defined in terms of a general strongly convex function. |
format |
text |
author |
WU, Liang LEDENT, Antoine LEI, Yunwen KLOFT, Marius |
author_facet |
WU, Liang LEDENT, Antoine LEI, Yunwen KLOFT, Marius |
author_sort |
WU, Liang |
title |
Fine-grained generalization analysis of vector-valued learning |
title_short |
Fine-grained generalization analysis of vector-valued learning |
title_full |
Fine-grained generalization analysis of vector-valued learning |
title_fullStr |
Fine-grained generalization analysis of vector-valued learning |
title_full_unstemmed |
Fine-grained generalization analysis of vector-valued learning |
title_sort |
fine-grained generalization analysis of vector-valued learning |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2021 |
url |
https://ink.library.smu.edu.sg/sis_research/7203 https://ink.library.smu.edu.sg/context/sis_research/article/8206/viewcontent/vector_val.pdf |
_version_ |
1770576269269794816 |