Bayesian neural network generalised additive models
In recent years, neural networks (NNs) have gained wide and lasting traction as the machine learning architecture of choice in many contexts, due to its flexibility and ability to represent complex functions. However, in the context of a regression task, NNs face difficulties in interpretability and...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/172098 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
Summary: | In recent years, neural networks (NNs) have gained wide and lasting traction as the machine learning architecture of choice in many contexts, due to its flexibility and ability to represent complex functions. However, in the context of a regression task, NNs face difficulties in interpretability and understanding of the effects of each predictor, due to the interactions between each predictor. Additive models, which are simpler models than NNs and lack interaction terms, allow insight into the effects of individual predictors, at the potential cost of model accuracy. More generally, machine learning models may also be ‘overconfident’ in their predictions; in that the model is unable to specify its confidence it is in its prediction. Taking a Bayesian viewpoint allows for machine learning models to represent its confidence (or lack thereof) in its predictions. This paper aims to
collect these ideas together to form a new machine learning architecture that is interpretable and
Bayesian in nature. |
---|