Decision trees, random vector functional link neural network and their ensembles
Decision Tree (DT) and Neural Network (NN) are two popular machine learning algorithms. A decision tree consists of internal nodes and leaf nodes. Each internal node is associated with a test function while each leaf node represents a class label or contains probability distribution. A decision tree...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Doctor of Philosophy |
Language: | English |
Published: |
Nanyang Technological University
2020
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/136779 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-136779 |
---|---|
record_format |
dspace |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering::Computing methodologies::Pattern recognition |
spellingShingle |
Engineering::Computer science and engineering::Computing methodologies::Pattern recognition Katuwal, Rakesh Kumar Decision trees, random vector functional link neural network and their ensembles |
description |
Decision Tree (DT) and Neural Network (NN) are two popular machine learning algorithms. A decision tree consists of internal nodes and leaf nodes. Each internal node is associated with a test function while each leaf node represents a class label or contains probability distribution. A decision tree is effectively a sequence of IF-THEN rules which is intuitive to humans and simple to implement. Meanwhile, a neural network is configured to mimic the functionality of the animal brain using highly interconnected system of computational units called neurons. The neurons are basically realized in the form of non-linear activation functions which process numeric inputs and output. The neurons are interconnected by transmission channels called synapse which is represented by the weights of the network connection. The weights are typically learned using back-propagation (BP) algorithm. The standard BP algorithm is time consuming, requires large amount of training data and may fail to converge. However, randomization based neural networks such as Random Vector Functional Link (RVFL) neural network eschew these issues by randomly fixing some parts (parameters) of the network while optimizing the rest using a closed form solution or an iterative procedure.
Ensemble of classifiers, also known as multiple classifier system, is a widely researched and frequently applied approach in machine learning. A multitude of studies corroborate that the combination of many classifiers leads to an improved performance compared to a single individual classifier. Decision trees and neural networks are typically used in an ensemble framework to obtain better performance.
Random Forest (RaF), an ensemble of decision trees, is extensively used two win several machine learning competitions and is considered one of the best algorithm. On the other hand, the popularity of Random Vector Functional Link (RVFL) neural network is also escalating because of its impressive performance in several diverse domains and its faster training time compared to conventional BP-trained NN. The first part of this thesis is based on random forest (including a hybrid ensemble classifier based on RaF and standard shallow RVFL) while the second part is based on multi-layer or deep random vector functional link neural network.
Decision trees in random forest usually employ binary splits at each node. The binary splits result in very deep trees. We utilize RVFL at the root node of the DTs to create multi-way splits. The proposed method provides a rich insight into the data by grouping the confusing or hard to classify samples and thus, provides an opportunity to employ fine-grained classification rule over the data. Also, by using multi-way splits, we obtain shallow trees that agree with Ockham\textquotesingle s razor principle of building smaller trees. Extensive experiments on several multi-class datasets show the efficacy of our method.
We also present our work in oblique decision trees. In oblique decision trees, an oblique (linear) hyperplane is employed instead of an axis-parallel (orthogonal) hyperplane at each non-leaf node to partition the data. Trees with such hyperplanes can better exploit the geometric structure in the data to increase the accuracy of the trees and reduce the depth. While the standard decision trees perform an exhaustive search for the best axis-parallel hyperplane, such an exhaustive search for the best oblique hyperplane is computationally expensive. Therefore, the present realizations of oblique decision trees do not evaluate many promising oblique splits to select the best. We present several oblique random forest variants that search for the best oblique hyperplane from a pool of homogeneous hyperplanes by employing one-vs-all approach. Similarly, we also propose a random forest of heterogeneous oblique decision trees that employ oblique splits generated via diverse linear classifiers at each non-leaf node on some top ranked partitions. In extensive experiments on 121 UCI machine learning datasets, 3 (out of 7) of our proposed oblique random forests take the top 3 ranks by outperforming other random forests and hundreds of other classifiers.
Deep learning, also known as representational learning, has sparked a surging interest in neural networks in recent years. Deep NN consists of several hidden layers stacked on top of each other wherein each hidden layer builds an internal representation of the data. With the current trend of building deeper networks, there have also been several attempts in the literature to build deep networks based on randomized neural networks. Even though there exist several deep learning models with randomized neural networks, there are limited works in the context of RVFL neural network. Thus, in this thesis, we propose several multi-layer (deep) random vector functional link neural networks variants. We also present an implicit ensemble of deep RVFL neural network which can be regarded as a marriage of ensemble and deep learning, that is simple and straight-forward to implement. |
author2 |
Ponnuthurai N. Suganthan |
author_facet |
Ponnuthurai N. Suganthan Katuwal, Rakesh Kumar |
format |
Thesis-Doctor of Philosophy |
author |
Katuwal, Rakesh Kumar |
author_sort |
Katuwal, Rakesh Kumar |
title |
Decision trees, random vector functional link neural network and their ensembles |
title_short |
Decision trees, random vector functional link neural network and their ensembles |
title_full |
Decision trees, random vector functional link neural network and their ensembles |
title_fullStr |
Decision trees, random vector functional link neural network and their ensembles |
title_full_unstemmed |
Decision trees, random vector functional link neural network and their ensembles |
title_sort |
decision trees, random vector functional link neural network and their ensembles |
publisher |
Nanyang Technological University |
publishDate |
2020 |
url |
https://hdl.handle.net/10356/136779 |
_version_ |
1772826313583755264 |
spelling |
sg-ntu-dr.10356-1367792023-07-04T17:21:38Z Decision trees, random vector functional link neural network and their ensembles Katuwal, Rakesh Kumar Ponnuthurai N. Suganthan School of Electrical and Electronic Engineering EPNSugan@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Pattern recognition Decision Tree (DT) and Neural Network (NN) are two popular machine learning algorithms. A decision tree consists of internal nodes and leaf nodes. Each internal node is associated with a test function while each leaf node represents a class label or contains probability distribution. A decision tree is effectively a sequence of IF-THEN rules which is intuitive to humans and simple to implement. Meanwhile, a neural network is configured to mimic the functionality of the animal brain using highly interconnected system of computational units called neurons. The neurons are basically realized in the form of non-linear activation functions which process numeric inputs and output. The neurons are interconnected by transmission channels called synapse which is represented by the weights of the network connection. The weights are typically learned using back-propagation (BP) algorithm. The standard BP algorithm is time consuming, requires large amount of training data and may fail to converge. However, randomization based neural networks such as Random Vector Functional Link (RVFL) neural network eschew these issues by randomly fixing some parts (parameters) of the network while optimizing the rest using a closed form solution or an iterative procedure. Ensemble of classifiers, also known as multiple classifier system, is a widely researched and frequently applied approach in machine learning. A multitude of studies corroborate that the combination of many classifiers leads to an improved performance compared to a single individual classifier. Decision trees and neural networks are typically used in an ensemble framework to obtain better performance. Random Forest (RaF), an ensemble of decision trees, is extensively used two win several machine learning competitions and is considered one of the best algorithm. On the other hand, the popularity of Random Vector Functional Link (RVFL) neural network is also escalating because of its impressive performance in several diverse domains and its faster training time compared to conventional BP-trained NN. The first part of this thesis is based on random forest (including a hybrid ensemble classifier based on RaF and standard shallow RVFL) while the second part is based on multi-layer or deep random vector functional link neural network. Decision trees in random forest usually employ binary splits at each node. The binary splits result in very deep trees. We utilize RVFL at the root node of the DTs to create multi-way splits. The proposed method provides a rich insight into the data by grouping the confusing or hard to classify samples and thus, provides an opportunity to employ fine-grained classification rule over the data. Also, by using multi-way splits, we obtain shallow trees that agree with Ockham\textquotesingle s razor principle of building smaller trees. Extensive experiments on several multi-class datasets show the efficacy of our method. We also present our work in oblique decision trees. In oblique decision trees, an oblique (linear) hyperplane is employed instead of an axis-parallel (orthogonal) hyperplane at each non-leaf node to partition the data. Trees with such hyperplanes can better exploit the geometric structure in the data to increase the accuracy of the trees and reduce the depth. While the standard decision trees perform an exhaustive search for the best axis-parallel hyperplane, such an exhaustive search for the best oblique hyperplane is computationally expensive. Therefore, the present realizations of oblique decision trees do not evaluate many promising oblique splits to select the best. We present several oblique random forest variants that search for the best oblique hyperplane from a pool of homogeneous hyperplanes by employing one-vs-all approach. Similarly, we also propose a random forest of heterogeneous oblique decision trees that employ oblique splits generated via diverse linear classifiers at each non-leaf node on some top ranked partitions. In extensive experiments on 121 UCI machine learning datasets, 3 (out of 7) of our proposed oblique random forests take the top 3 ranks by outperforming other random forests and hundreds of other classifiers. Deep learning, also known as representational learning, has sparked a surging interest in neural networks in recent years. Deep NN consists of several hidden layers stacked on top of each other wherein each hidden layer builds an internal representation of the data. With the current trend of building deeper networks, there have also been several attempts in the literature to build deep networks based on randomized neural networks. Even though there exist several deep learning models with randomized neural networks, there are limited works in the context of RVFL neural network. Thus, in this thesis, we propose several multi-layer (deep) random vector functional link neural networks variants. We also present an implicit ensemble of deep RVFL neural network which can be regarded as a marriage of ensemble and deep learning, that is simple and straight-forward to implement. Doctor of Philosophy 2020-01-24T03:24:21Z 2020-01-24T03:24:21Z 2019 Thesis-Doctor of Philosophy Katuwal, R. K. (2019). Decision trees, random vector functional link neural network and their ensembles. Doctoral thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/136779 10.32657/10356/136779 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University |