Gradient boosting with piece-wise linear regression trees

Gradient Boosted Decision Trees (GBDT) is a very successful ensemble learning algorithm widely used across a variety of applications. Recently, several variants of GBDT training algorithms and implementations have been designed and heavily optimized in some very popular open sourced toolkits includi...

Full description

Saved in:
Bibliographic Details
Main Authors: SHI, Yu, LI, Jian, LI, Zhize
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2019
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8675
https://ink.library.smu.edu.sg/context/sis_research/article/9678/viewcontent/IJCAI19_full_GBDT.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9678
record_format dspace
spelling sg-smu-ink.sis_research-96782024-03-28T09:07:59Z Gradient boosting with piece-wise linear regression trees SHI, Yu LI, Jian LI, Zhize Gradient Boosted Decision Trees (GBDT) is a very successful ensemble learning algorithm widely used across a variety of applications. Recently, several variants of GBDT training algorithms and implementations have been designed and heavily optimized in some very popular open sourced toolkits including XGBoost, LightGBM and CatBoost. In this paper, we show that both the accuracy and efficiency of GBDT can be further enhanced by using more complex base learners. Specifically, we extend gradient boosting to use piecewise linear regression trees (PL Trees), instead of piecewise constant regression trees, as base learners. We show that PL Trees can accelerate convergence of GBDT and improve the accuracy. We also propose some optimization tricks to substantially reduce the training time of PL Trees, with little sacrifice of accuracy. Moreover, we propose several implementation techniques to speedup our algorithm on modern computer architectures with powerful Single Instruction Multiple Data (SIMD) parallelism. The experimental results show that GBDT with PL Trees can provide very competitive testing accuracy with comparable or less training time. 2019-08-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8675 https://ink.library.smu.edu.sg/context/sis_research/article/9678/viewcontent/IJCAI19_full_GBDT.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Databases and Information Systems
spellingShingle Databases and Information Systems
SHI, Yu
LI, Jian
LI, Zhize
Gradient boosting with piece-wise linear regression trees
description Gradient Boosted Decision Trees (GBDT) is a very successful ensemble learning algorithm widely used across a variety of applications. Recently, several variants of GBDT training algorithms and implementations have been designed and heavily optimized in some very popular open sourced toolkits including XGBoost, LightGBM and CatBoost. In this paper, we show that both the accuracy and efficiency of GBDT can be further enhanced by using more complex base learners. Specifically, we extend gradient boosting to use piecewise linear regression trees (PL Trees), instead of piecewise constant regression trees, as base learners. We show that PL Trees can accelerate convergence of GBDT and improve the accuracy. We also propose some optimization tricks to substantially reduce the training time of PL Trees, with little sacrifice of accuracy. Moreover, we propose several implementation techniques to speedup our algorithm on modern computer architectures with powerful Single Instruction Multiple Data (SIMD) parallelism. The experimental results show that GBDT with PL Trees can provide very competitive testing accuracy with comparable or less training time.
format text
author SHI, Yu
LI, Jian
LI, Zhize
author_facet SHI, Yu
LI, Jian
LI, Zhize
author_sort SHI, Yu
title Gradient boosting with piece-wise linear regression trees
title_short Gradient boosting with piece-wise linear regression trees
title_full Gradient boosting with piece-wise linear regression trees
title_fullStr Gradient boosting with piece-wise linear regression trees
title_full_unstemmed Gradient boosting with piece-wise linear regression trees
title_sort gradient boosting with piece-wise linear regression trees
publisher Institutional Knowledge at Singapore Management University
publishDate 2019
url https://ink.library.smu.edu.sg/sis_research/8675
https://ink.library.smu.edu.sg/context/sis_research/article/9678/viewcontent/IJCAI19_full_GBDT.pdf
_version_ 1795302169499402240