Fine-grained generalization analysis of vector-valued learning

Many fundamental machine learning tasks can be formulated as a problem of learning with vector-valued functions, where we learn multiple scalar-valued functions together. Although there is some generalization analysis on different specific algorithms under the empirical risk minimization principle,...

Full description

Saved in:
Bibliographic Details
Main Authors: WU, Liang, LEDENT, Antoine, LEI, Yunwen, KLOFT, Marius
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2021
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7203
https://ink.library.smu.edu.sg/context/sis_research/article/8206/viewcontent/vector_val.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8206
record_format dspace
spelling sg-smu-ink.sis_research-82062022-08-04T08:49:28Z Fine-grained generalization analysis of vector-valued learning WU, Liang LEDENT, Antoine LEI, Yunwen KLOFT, Marius Many fundamental machine learning tasks can be formulated as a problem of learning with vector-valued functions, where we learn multiple scalar-valued functions together. Although there is some generalization analysis on different specific algorithms under the empirical risk minimization principle, a unifying analysis of vector-valued learning under a regularization framework is still lacking. In this paper, we initiate the generalization analysis of regularized vector-valued learning algorithms by presenting bounds with a mild dependency on the output dimension and a fast rate on the sample size. Our discussions relax the existing assumptions on the restrictive constraint of hypothesis spaces, smoothness of loss functions and low-noise condition. To understand the interaction between optimization and learning, we further use our results to derive the first generalization bounds for stochastic gradient descent with vector-valued functions. We apply our general results to multi-class classification and multi-label classification, which yield the first bounds with a logarithmic dependency on the output dimension for extreme multi-label classification with the Frobenius regularization. As a byproduct, we derive a Rademacher complexity bound for loss function classes defined in terms of a general strongly convex function. 2021-02-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7203 https://ink.library.smu.edu.sg/context/sis_research/article/8206/viewcontent/vector_val.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Statistical Learning Theory Multi-label Learning Stochastic Gradient Descent Artificial Intelligence and Robotics Theory and Algorithms
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Statistical Learning Theory
Multi-label Learning
Stochastic Gradient Descent
Artificial Intelligence and Robotics
Theory and Algorithms
spellingShingle Statistical Learning Theory
Multi-label Learning
Stochastic Gradient Descent
Artificial Intelligence and Robotics
Theory and Algorithms
WU, Liang
LEDENT, Antoine
LEI, Yunwen
KLOFT, Marius
Fine-grained generalization analysis of vector-valued learning
description Many fundamental machine learning tasks can be formulated as a problem of learning with vector-valued functions, where we learn multiple scalar-valued functions together. Although there is some generalization analysis on different specific algorithms under the empirical risk minimization principle, a unifying analysis of vector-valued learning under a regularization framework is still lacking. In this paper, we initiate the generalization analysis of regularized vector-valued learning algorithms by presenting bounds with a mild dependency on the output dimension and a fast rate on the sample size. Our discussions relax the existing assumptions on the restrictive constraint of hypothesis spaces, smoothness of loss functions and low-noise condition. To understand the interaction between optimization and learning, we further use our results to derive the first generalization bounds for stochastic gradient descent with vector-valued functions. We apply our general results to multi-class classification and multi-label classification, which yield the first bounds with a logarithmic dependency on the output dimension for extreme multi-label classification with the Frobenius regularization. As a byproduct, we derive a Rademacher complexity bound for loss function classes defined in terms of a general strongly convex function.
format text
author WU, Liang
LEDENT, Antoine
LEI, Yunwen
KLOFT, Marius
author_facet WU, Liang
LEDENT, Antoine
LEI, Yunwen
KLOFT, Marius
author_sort WU, Liang
title Fine-grained generalization analysis of vector-valued learning
title_short Fine-grained generalization analysis of vector-valued learning
title_full Fine-grained generalization analysis of vector-valued learning
title_fullStr Fine-grained generalization analysis of vector-valued learning
title_full_unstemmed Fine-grained generalization analysis of vector-valued learning
title_sort fine-grained generalization analysis of vector-valued learning
publisher Institutional Knowledge at Singapore Management University
publishDate 2021
url https://ink.library.smu.edu.sg/sis_research/7203
https://ink.library.smu.edu.sg/context/sis_research/article/8206/viewcontent/vector_val.pdf
_version_ 1770576269269794816