Efficient stochastic gradient hard thresholding

Stochastic gradient hard thresholding methods have recently been shown to work favorably in solving large-scale empirical risk minimization problems under sparsity or rank constraint. Despite the improved iteration complexity over full gradient methods, the gradient evaluation and hard thresholding...

Full description

Saved in:
Bibliographic Details
Main Authors: ZHOU, Pan, YUAN, Xiao-Tong, FENG, Jiashi
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2018
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/9003
https://ink.library.smu.edu.sg/context/sis_research/article/10006/viewcontent/NeurIPS_2018_efficient_stochastic_gradient_hard_thresholding_Paper.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-10006
record_format dspace
spelling sg-smu-ink.sis_research-100062024-07-25T08:17:01Z Efficient stochastic gradient hard thresholding ZHOU, Pan YUAN, Xiao-Tong FENG, Jiashi Stochastic gradient hard thresholding methods have recently been shown to work favorably in solving large-scale empirical risk minimization problems under sparsity or rank constraint. Despite the improved iteration complexity over full gradient methods, the gradient evaluation and hard thresholding complexity of the existing stochastic algorithms usually scales linearly with data size, which could still be expensive when data is huge and the hard thresholding step could be as expensive as singular value decomposition in rank-constrained problems. To address these deficiencies, we propose an efficient hybrid stochastic gradient hard thresholding (HSG-HT) method that can be provably shown to have sample-size-independent gradient evaluation and hard thresholding complexity bounds. Specifically, we prove that the stochastic gradient evaluation complexity of HSG-HT scales linearly with inverse of sub-optimality and its hard thresholding complexity scales logarithmically. By applying the heavy ball acceleration technique, we further propose an accelerated variant of HSG-HT which can be shown to have improved factor dependence on restricted condition number in the quadratic case. Numerical results confirm our theoretical affirmation and demonstrate the computational efficiency of the proposed methods. 2018-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9003 https://ink.library.smu.edu.sg/context/sis_research/article/10006/viewcontent/NeurIPS_2018_efficient_stochastic_gradient_hard_thresholding_Paper.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University OS and Networks Theory and Algorithms
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic OS and Networks
Theory and Algorithms
spellingShingle OS and Networks
Theory and Algorithms
ZHOU, Pan
YUAN, Xiao-Tong
FENG, Jiashi
Efficient stochastic gradient hard thresholding
description Stochastic gradient hard thresholding methods have recently been shown to work favorably in solving large-scale empirical risk minimization problems under sparsity or rank constraint. Despite the improved iteration complexity over full gradient methods, the gradient evaluation and hard thresholding complexity of the existing stochastic algorithms usually scales linearly with data size, which could still be expensive when data is huge and the hard thresholding step could be as expensive as singular value decomposition in rank-constrained problems. To address these deficiencies, we propose an efficient hybrid stochastic gradient hard thresholding (HSG-HT) method that can be provably shown to have sample-size-independent gradient evaluation and hard thresholding complexity bounds. Specifically, we prove that the stochastic gradient evaluation complexity of HSG-HT scales linearly with inverse of sub-optimality and its hard thresholding complexity scales logarithmically. By applying the heavy ball acceleration technique, we further propose an accelerated variant of HSG-HT which can be shown to have improved factor dependence on restricted condition number in the quadratic case. Numerical results confirm our theoretical affirmation and demonstrate the computational efficiency of the proposed methods.
format text
author ZHOU, Pan
YUAN, Xiao-Tong
FENG, Jiashi
author_facet ZHOU, Pan
YUAN, Xiao-Tong
FENG, Jiashi
author_sort ZHOU, Pan
title Efficient stochastic gradient hard thresholding
title_short Efficient stochastic gradient hard thresholding
title_full Efficient stochastic gradient hard thresholding
title_fullStr Efficient stochastic gradient hard thresholding
title_full_unstemmed Efficient stochastic gradient hard thresholding
title_sort efficient stochastic gradient hard thresholding
publisher Institutional Knowledge at Singapore Management University
publishDate 2018
url https://ink.library.smu.edu.sg/sis_research/9003
https://ink.library.smu.edu.sg/context/sis_research/article/10006/viewcontent/NeurIPS_2018_efficient_stochastic_gradient_hard_thresholding_Paper.pdf
_version_ 1814047689312043008