Stochastic gradient Hamiltonian Monte Carlo with variance reduction for Bayesian inference
Gradient-based Monte Carlo sampling algorithms, like Langevin dynamics and Hamiltonian Monte Carlo, are important methods for Bayesian inference. In large-scale settings, full-gradients are not affordable and thus stochastic gradients evaluated on mini-batches are used as a replacement. In order to...
Saved in:
Main Authors: | , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2019
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8689 https://ink.library.smu.edu.sg/context/sis_research/article/9692/viewcontent/ML19_vrhmc.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9692 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-96922024-03-28T08:45:18Z Stochastic gradient Hamiltonian Monte Carlo with variance reduction for Bayesian inference LI, Zhize ZHANG, Tianyi CHENG, Shuyu ZHU, Jun LI, Jian Gradient-based Monte Carlo sampling algorithms, like Langevin dynamics and Hamiltonian Monte Carlo, are important methods for Bayesian inference. In large-scale settings, full-gradients are not affordable and thus stochastic gradients evaluated on mini-batches are used as a replacement. In order to reduce the high variance of noisy stochastic gradients, Dubey et al. (in: Advances in neural information processing systems, pp 1154–1162, 2016) applied the standard variance reduction technique on stochastic gradient Langevin dynamics and obtained both theoretical and experimental improvements. In this paper, we apply the variance reduction tricks on Hamiltonian Monte Carlo and achieve better theoretical convergence results compared with the variance-reduced Langevin dynamics. Moreover, we apply the symmetric splitting scheme in our variance-reduced Hamiltonian Monte Carlo algorithms to further improve the theoretical results. The experimental results are also consistent with the theoretical results. As our experiment shows, variance-reduced Hamiltonian Monte Carlo demonstrates better performance than variance-reduced Langevin dynamics in Bayesian regression and classification tasks on real-world datasets. 2019-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8689 info:doi/10.1007/s10994-019-05825-y https://ink.library.smu.edu.sg/context/sis_research/article/9692/viewcontent/ML19_vrhmc.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Hamiltonian Monte Carlo Variance reduction Bayesian inference Databases and Information Systems |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Hamiltonian Monte Carlo Variance reduction Bayesian inference Databases and Information Systems |
spellingShingle |
Hamiltonian Monte Carlo Variance reduction Bayesian inference Databases and Information Systems LI, Zhize ZHANG, Tianyi CHENG, Shuyu ZHU, Jun LI, Jian Stochastic gradient Hamiltonian Monte Carlo with variance reduction for Bayesian inference |
description |
Gradient-based Monte Carlo sampling algorithms, like Langevin dynamics and Hamiltonian Monte Carlo, are important methods for Bayesian inference. In large-scale settings, full-gradients are not affordable and thus stochastic gradients evaluated on mini-batches are used as a replacement. In order to reduce the high variance of noisy stochastic gradients, Dubey et al. (in: Advances in neural information processing systems, pp 1154–1162, 2016) applied the standard variance reduction technique on stochastic gradient Langevin dynamics and obtained both theoretical and experimental improvements. In this paper, we apply the variance reduction tricks on Hamiltonian Monte Carlo and achieve better theoretical convergence results compared with the variance-reduced Langevin dynamics. Moreover, we apply the symmetric splitting scheme in our variance-reduced Hamiltonian Monte Carlo algorithms to further improve the theoretical results. The experimental results are also consistent with the theoretical results. As our experiment shows, variance-reduced Hamiltonian Monte Carlo demonstrates better performance than variance-reduced Langevin dynamics in Bayesian regression and classification tasks on real-world datasets. |
format |
text |
author |
LI, Zhize ZHANG, Tianyi CHENG, Shuyu ZHU, Jun LI, Jian |
author_facet |
LI, Zhize ZHANG, Tianyi CHENG, Shuyu ZHU, Jun LI, Jian |
author_sort |
LI, Zhize |
title |
Stochastic gradient Hamiltonian Monte Carlo with variance reduction for Bayesian inference |
title_short |
Stochastic gradient Hamiltonian Monte Carlo with variance reduction for Bayesian inference |
title_full |
Stochastic gradient Hamiltonian Monte Carlo with variance reduction for Bayesian inference |
title_fullStr |
Stochastic gradient Hamiltonian Monte Carlo with variance reduction for Bayesian inference |
title_full_unstemmed |
Stochastic gradient Hamiltonian Monte Carlo with variance reduction for Bayesian inference |
title_sort |
stochastic gradient hamiltonian monte carlo with variance reduction for bayesian inference |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2019 |
url |
https://ink.library.smu.edu.sg/sis_research/8689 https://ink.library.smu.edu.sg/context/sis_research/article/9692/viewcontent/ML19_vrhmc.pdf |
_version_ |
1795302173776543744 |