Neural network training and rule extraction with augmented discretized input

© 2016 Elsevier B.V. The classification and prediction accuracy of neural networks can be improved when they are trained with discretized continuous attributes as additional inputs. Such input augmentation makes it easier for the network weights to form more accurate decision boundaries when the dat...

Full description

Saved in:
Bibliographic Details
Main Authors: Hayashi, Yoichi, Setiono, Rudy, Azcarraga, Arnulfo P.
Format: text
Published: Animo Repository 2016
Online Access:https://animorepository.dlsu.edu.ph/faculty_research/931
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: De La Salle University
id oai:animorepository.dlsu.edu.ph:faculty_research-1930
record_format eprints
spelling oai:animorepository.dlsu.edu.ph:faculty_research-19302022-11-16T02:39:51Z Neural network training and rule extraction with augmented discretized input Hayashi, Yoichi Setiono, Rudy Azcarraga, Arnulfo P. © 2016 Elsevier B.V. The classification and prediction accuracy of neural networks can be improved when they are trained with discretized continuous attributes as additional inputs. Such input augmentation makes it easier for the network weights to form more accurate decision boundaries when the data samples of different classes in the data set are contained in distinct hyper-rectangular subregions in the original input space. In this paper, we present first how a neural network can be trained with augmented discretized inputs. The additional inputs are obtained by dividing the original interval of each continuous attribute into subintervals of equal length. The network is then pruned to remove most of the discretized inputs as well as the original continuous attributes as long as the network still achieves a minimum preset accuracy requirement. We then discuss how comprehensible classification rules can be extracted from the pruned network by analyzing the activations of the network hidden units and the weights of the network connections that remain in the pruned network. Our experiments on artificial data sets show that the rules extracted from the neural networks can perfectly replicate the class membership rules used to create the data perfectly. On real-life benchmark data sets, neural networks trained with augmented discretized inputs are shown to achieve better accuracy than neural networks trained with the original data. 2016-09-26T07:00:00Z text text/html https://animorepository.dlsu.edu.ph/faculty_research/931 Faculty Research Work Animo Repository
institution De La Salle University
building De La Salle University Library
continent Asia
country Philippines
Philippines
content_provider De La Salle University Library
collection DLSU Institutional Repository
description © 2016 Elsevier B.V. The classification and prediction accuracy of neural networks can be improved when they are trained with discretized continuous attributes as additional inputs. Such input augmentation makes it easier for the network weights to form more accurate decision boundaries when the data samples of different classes in the data set are contained in distinct hyper-rectangular subregions in the original input space. In this paper, we present first how a neural network can be trained with augmented discretized inputs. The additional inputs are obtained by dividing the original interval of each continuous attribute into subintervals of equal length. The network is then pruned to remove most of the discretized inputs as well as the original continuous attributes as long as the network still achieves a minimum preset accuracy requirement. We then discuss how comprehensible classification rules can be extracted from the pruned network by analyzing the activations of the network hidden units and the weights of the network connections that remain in the pruned network. Our experiments on artificial data sets show that the rules extracted from the neural networks can perfectly replicate the class membership rules used to create the data perfectly. On real-life benchmark data sets, neural networks trained with augmented discretized inputs are shown to achieve better accuracy than neural networks trained with the original data.
format text
author Hayashi, Yoichi
Setiono, Rudy
Azcarraga, Arnulfo P.
spellingShingle Hayashi, Yoichi
Setiono, Rudy
Azcarraga, Arnulfo P.
Neural network training and rule extraction with augmented discretized input
author_facet Hayashi, Yoichi
Setiono, Rudy
Azcarraga, Arnulfo P.
author_sort Hayashi, Yoichi
title Neural network training and rule extraction with augmented discretized input
title_short Neural network training and rule extraction with augmented discretized input
title_full Neural network training and rule extraction with augmented discretized input
title_fullStr Neural network training and rule extraction with augmented discretized input
title_full_unstemmed Neural network training and rule extraction with augmented discretized input
title_sort neural network training and rule extraction with augmented discretized input
publisher Animo Repository
publishDate 2016
url https://animorepository.dlsu.edu.ph/faculty_research/931
_version_ 1751550429709205504