Large-scale heteroscedastic regression via Gaussian process
Heteroscedastic regression considering the varying noises among observations has many applications in the fields, such as machine learning and statistics. Here, we focus on the heteroscedastic Gaussian process (HGP) regression that integrates the latent function and the noise function in a unified n...
Saved in:
Main Authors: | , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/159646 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-159646 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1596462022-06-28T08:43:06Z Large-scale heteroscedastic regression via Gaussian process Liu, Haitao Ong, Yew-Soon Cai, Jianfei School of Computer Science and Engineering Rolls-Royce@NTU Corporate Lab Engineering::Computer science and engineering Distributed Learning Sparse Approximation Heteroscedastic regression considering the varying noises among observations has many applications in the fields, such as machine learning and statistics. Here, we focus on the heteroscedastic Gaussian process (HGP) regression that integrates the latent function and the noise function in a unified nonparametric Bayesian framework. Though showing remarkable performance, HGP suffers from the cubic time complexity, which strictly limits its application to big data. To improve the scalability, we first develop a variational sparse inference algorithm, named VSHGP, to handle large-scale data sets. Furthermore, two variants are developed to improve the scalability and capability of VSHGP. The first is stochastic VSHGP (SVSHGP) that derives a factorized evidence lower bound, thus enhancing efficient stochastic variational inference. The second is distributed VSHGP (DVSHGP) that follows the Bayesian committee machine formalism to distribute computations over multiple local VSHGP experts with many inducing points and adopts hybrid parameters for experts to guard against overfitting and capture local variety. The superiority of DVSHGP and SVSHGP compared to the existing scalable HGP/homoscedastic GP is then extensively verified on various data sets. Nanyang Technological University National Research Foundation (NRF) This work was supported in part by the National Research Foundation, Singapore, under its AI Singapore Program under Award AISG-RP-2018-004, in part by the Data Science and Artificial Intelligence Research Center (DSAIR), Nanyang Technological University, and in part by the Rolls-Royce@NTU Corporate Laboratory. 2022-06-28T08:43:06Z 2022-06-28T08:43:06Z 2020 Journal Article Liu, H., Ong, Y. & Cai, J. (2020). Large-scale heteroscedastic regression via Gaussian process. IEEE Transactions On Neural Networks and Learning Systems, 32(2), 708-721. https://dx.doi.org/10.1109/TNNLS.2020.2979188 2162-237X https://hdl.handle.net/10356/159646 10.1109/TNNLS.2020.2979188 32275610 2-s2.0-85100738494 2 32 708 721 en AISG-RP-2018-004 IEEE Transactions on Neural Networks and Learning Systems © 2020 IEEE. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Distributed Learning Sparse Approximation |
spellingShingle |
Engineering::Computer science and engineering Distributed Learning Sparse Approximation Liu, Haitao Ong, Yew-Soon Cai, Jianfei Large-scale heteroscedastic regression via Gaussian process |
description |
Heteroscedastic regression considering the varying noises among observations has many applications in the fields, such as machine learning and statistics. Here, we focus on the heteroscedastic Gaussian process (HGP) regression that integrates the latent function and the noise function in a unified nonparametric Bayesian framework. Though showing remarkable performance, HGP suffers from the cubic time complexity, which strictly limits its application to big data. To improve the scalability, we first develop a variational sparse inference algorithm, named VSHGP, to handle large-scale data sets. Furthermore, two variants are developed to improve the scalability and capability of VSHGP. The first is stochastic VSHGP (SVSHGP) that derives a factorized evidence lower bound, thus enhancing efficient stochastic variational inference. The second is distributed VSHGP (DVSHGP) that follows the Bayesian committee machine formalism to distribute computations over multiple local VSHGP experts with many inducing points and adopts hybrid parameters for experts to guard against overfitting and capture local variety. The superiority of DVSHGP and SVSHGP compared to the existing scalable HGP/homoscedastic GP is then extensively verified on various data sets. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Liu, Haitao Ong, Yew-Soon Cai, Jianfei |
format |
Article |
author |
Liu, Haitao Ong, Yew-Soon Cai, Jianfei |
author_sort |
Liu, Haitao |
title |
Large-scale heteroscedastic regression via Gaussian process |
title_short |
Large-scale heteroscedastic regression via Gaussian process |
title_full |
Large-scale heteroscedastic regression via Gaussian process |
title_fullStr |
Large-scale heteroscedastic regression via Gaussian process |
title_full_unstemmed |
Large-scale heteroscedastic regression via Gaussian process |
title_sort |
large-scale heteroscedastic regression via gaussian process |
publishDate |
2022 |
url |
https://hdl.handle.net/10356/159646 |
_version_ |
1738844823667867648 |