Two-layer EM algorithm for ALD mixture regression models: A new solution to composite quantile regression
We advocate linear regression by modeling the error term through a finite mixture of asymmetric Laplace distributions (ALDs). The model expands the flexibility of linear regression to account for heterogeneity among data and allows us to establish the equivalence between maximum likelihood estimatio...
Saved in:
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2017
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/83377 http://hdl.handle.net/10220/43534 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-83377 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-833772023-02-28T19:32:50Z Two-layer EM algorithm for ALD mixture regression models: A new solution to composite quantile regression Wang, Shangshan Xiang, Liming School of Physical and Mathematical Sciences Asymmetric Laplace distribution Composite quantile regression We advocate linear regression by modeling the error term through a finite mixture of asymmetric Laplace distributions (ALDs). The model expands the flexibility of linear regression to account for heterogeneity among data and allows us to establish the equivalence between maximum likelihood estimation of the model parameters and the composite quantile regression (CQR) estimation developed by Zou and Yuan (Ann. Stat. 36:1108–1126, 2008), providing a new likelihood-based solution to CQR. Particularly, we develop a computationally efficient estimation procedure via a two-layer EM algorithm, where the first layer EM algorithm incorporates missing information from the component memberships of the mixture model and nests the second layer EM in its M-step to accommodate latent variables involved in the location-scale mixture representation of the ALD. An appealing feature of the proposed algorithm is that the closed form updates for parameters in each iteration are obtained explicitly, instead of resorting to numerical optimization methods as in the existing work. Computational complexity can be reduced significantly. We evaluate the performance through simulation studies and illustrate its usefulness by analyzing a gene expression dataset. MOE (Min. of Education, S’pore) Accepted version 2017-08-03T06:56:12Z 2019-12-06T15:21:07Z 2017-08-03T06:56:12Z 2019-12-06T15:21:07Z 2017 Journal Article Wang, S., & Xiang, L. (2017). Two-layer EM algorithm for ALD mixture regression models: A new solution to composite quantile regression. Computational Statistics & Data Analysis, 115, 136-154. 0167-9473 https://hdl.handle.net/10356/83377 http://hdl.handle.net/10220/43534 10.1016/j.csda.2017.06.002 en Computational Statistics & Data Analysis © 2017 Elsevier. This is the author created version of a work that has been peer reviewed and accepted for publication by Computational Statistics & Data Analysis, Elsevier. It incorporates referee’s comments but changes resulting from the publishing process, such as copyediting, structural formatting, may not be reflected in this document. The published version is available at: [http://dx.doi.org/10.1016/j.csda.2017.06.002]. 33 p. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Asymmetric Laplace distribution Composite quantile regression |
spellingShingle |
Asymmetric Laplace distribution Composite quantile regression Wang, Shangshan Xiang, Liming Two-layer EM algorithm for ALD mixture regression models: A new solution to composite quantile regression |
description |
We advocate linear regression by modeling the error term through a finite mixture of asymmetric Laplace distributions (ALDs). The model expands the flexibility of linear regression to account for heterogeneity among data and allows us to establish the equivalence between maximum likelihood estimation of the model parameters and the composite quantile regression (CQR) estimation developed by Zou and Yuan (Ann. Stat. 36:1108–1126, 2008), providing a new likelihood-based solution to CQR. Particularly, we develop a computationally efficient estimation procedure via a two-layer EM algorithm, where the first layer EM algorithm incorporates missing information from the component memberships of the mixture model and nests the second layer EM in its M-step to accommodate latent variables involved in the location-scale mixture representation of the ALD. An appealing feature of the proposed algorithm is that the closed form updates for parameters in each iteration are obtained explicitly, instead of resorting to numerical optimization methods as in the existing work. Computational complexity can be reduced significantly. We evaluate the performance through simulation studies and illustrate its usefulness by analyzing a gene expression dataset. |
author2 |
School of Physical and Mathematical Sciences |
author_facet |
School of Physical and Mathematical Sciences Wang, Shangshan Xiang, Liming |
format |
Article |
author |
Wang, Shangshan Xiang, Liming |
author_sort |
Wang, Shangshan |
title |
Two-layer EM algorithm for ALD mixture regression models: A new solution to composite quantile regression |
title_short |
Two-layer EM algorithm for ALD mixture regression models: A new solution to composite quantile regression |
title_full |
Two-layer EM algorithm for ALD mixture regression models: A new solution to composite quantile regression |
title_fullStr |
Two-layer EM algorithm for ALD mixture regression models: A new solution to composite quantile regression |
title_full_unstemmed |
Two-layer EM algorithm for ALD mixture regression models: A new solution to composite quantile regression |
title_sort |
two-layer em algorithm for ald mixture regression models: a new solution to composite quantile regression |
publishDate |
2017 |
url |
https://hdl.handle.net/10356/83377 http://hdl.handle.net/10220/43534 |
_version_ |
1759853479490224128 |