Deep learning architecture analysis with mapper
In recent years, we have seen the rise of deep learning models such as convolutional neural networks (CNN) for image classification. However, we do not understand what makes these networks able to achieve such outstanding performance. Upon building a neural network, we can only see the input and out...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/166418 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-166418 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1664182023-05-01T15:35:56Z Deep learning architecture analysis with mapper Foo, Kelvin Moo Chen Xia Kelin School of Physical and Mathematical Sciences xiakelin@ntu.edu.sg Science::Mathematics::Topology In recent years, we have seen the rise of deep learning models such as convolutional neural networks (CNN) for image classification. However, we do not understand what makes these networks able to achieve such outstanding performance. Upon building a neural network, we can only see the input and output of the model and the in-between remains as a mystery or a black box to us. In this paper, given a trained deep neural network, we address the interpretability issue by probing neuron activation. We use a tool in topological data analysis (TDA), known as mapper, to visualize relationships between different activation in a particular layer of the specified neural network. Mapper provides two topological summaries, namely branches and loops. The effectiveness of mapper depends on the dataset being used. In the case of image classification tasks, if the images are dissimilar, mapper can construct informative branches to visualize the relationships between activation. However, if the images are very similar, mapper is not useful. For tabular data, mapper is useful only if the majority of the features are continuous variables, as demonstrated by the Iris dataset example. For text data, the usefulness of mapper in visualizing the activation is determined by the length and content of the text. If the text is short and focuses on the same content, mapper is not useful for visualizing the activation. Bachelor of Science in Mathematical Sciences 2023-04-26T05:27:21Z 2023-04-26T05:27:21Z 2023 Final Year Project (FYP) Foo, K. M. C. (2023). Deep learning architecture analysis with mapper. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/166418 https://hdl.handle.net/10356/166418 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Science::Mathematics::Topology |
spellingShingle |
Science::Mathematics::Topology Foo, Kelvin Moo Chen Deep learning architecture analysis with mapper |
description |
In recent years, we have seen the rise of deep learning models such as convolutional neural networks (CNN) for image classification. However, we do not understand what makes these networks able to achieve such outstanding performance. Upon building a neural network, we can only see the input and output of the model and the in-between remains as a mystery or a black box to us. In this paper, given a trained deep neural network, we address the interpretability issue by probing neuron activation. We use a tool in topological data analysis (TDA), known as mapper, to visualize relationships between different activation in a particular layer of the specified neural network. Mapper provides two topological summaries, namely branches and loops. The effectiveness of mapper depends on the dataset being used. In the case of image classification tasks, if the images are dissimilar, mapper can construct informative branches to visualize the relationships between activation. However, if the images are very similar, mapper is not useful. For tabular data, mapper is useful only if the majority of the features are continuous variables, as demonstrated by the Iris dataset example. For text data, the usefulness of mapper in visualizing the activation is determined by the length and content of the text. If the text is short and focuses on the same content, mapper is not useful for visualizing the activation. |
author2 |
Xia Kelin |
author_facet |
Xia Kelin Foo, Kelvin Moo Chen |
format |
Final Year Project |
author |
Foo, Kelvin Moo Chen |
author_sort |
Foo, Kelvin Moo Chen |
title |
Deep learning architecture analysis with mapper |
title_short |
Deep learning architecture analysis with mapper |
title_full |
Deep learning architecture analysis with mapper |
title_fullStr |
Deep learning architecture analysis with mapper |
title_full_unstemmed |
Deep learning architecture analysis with mapper |
title_sort |
deep learning architecture analysis with mapper |
publisher |
Nanyang Technological University |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/166418 |
_version_ |
1765213842125619200 |