Interpretable vector language models
Natural Language Processing (NLP) is an important part of Artificial Intelligence (AI) that aims to create algorithms which improve how humans understand and interpret bodies of text. In particular, word embeddings form a vital part of NLP, as models like Word2Vec and GloVe assign numeric vectors to...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/175573 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-175573 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1755732024-05-06T15:37:34Z Interpretable vector language models Siow, Zi Hao Fedor Duzhin School of Physical and Mathematical Sciences FDuzhin@ntu.edu.sg Mathematical Sciences Interpretable Vector language models Data visualisation Natural Language Processing (NLP) is an important part of Artificial Intelligence (AI) that aims to create algorithms which improve how humans understand and interpret bodies of text. In particular, word embeddings form a vital part of NLP, as models like Word2Vec and GloVe assign numeric vectors to words in a text corpus such that norms and angles between words are preserved and semantic structure is maintained. While their effectiveness is undisputed, they face a major limitation in the form of limited interpretability, as individual entries are hard to interpret due to the simultaneous rotation of all vectors preserving semantic structure while entries become mixed up. Hence, in this study, we proposed a novel approach of generating word embeddings with a higher degree of interpretability. We associated the interpretability of a word embedding with the optimisation of various loss functions, namely Varimax, Quartimax and the l1-norm, defined on the Lie group SO(d). Our findings revealed that the l1-norm method achieved the highest level of interpretability among the three methods, because its solutions tend to have a higher proportion of matrix elements that are close to zero by promoting sparsity. Through this study, we hope to have provided valuable insights into creating word embeddings with more interpretable entries. Bachelor's degree 2024-04-30T03:57:24Z 2024-04-30T03:57:24Z 2024 Final Year Project (FYP) Siow, Z. H. (2024). Interpretable vector language models. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/175573 https://hdl.handle.net/10356/175573 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Mathematical Sciences Interpretable Vector language models Data visualisation |
spellingShingle |
Mathematical Sciences Interpretable Vector language models Data visualisation Siow, Zi Hao Interpretable vector language models |
description |
Natural Language Processing (NLP) is an important part of Artificial Intelligence (AI) that aims to create algorithms which improve how humans understand and interpret bodies of text. In particular, word embeddings form a vital part of NLP, as models like Word2Vec and GloVe assign numeric vectors to words in a text corpus such that norms and angles between words are preserved and semantic structure is maintained. While their effectiveness is undisputed, they face a major limitation in the form of limited interpretability, as individual entries are hard to interpret due to the simultaneous rotation of all vectors preserving semantic structure while entries become mixed up. Hence, in this study, we proposed a novel approach of generating word embeddings with a higher degree of interpretability. We associated the interpretability of a word embedding with the optimisation of various loss functions, namely Varimax, Quartimax and the l1-norm, defined on the Lie group SO(d). Our findings revealed that the l1-norm method achieved the highest level of interpretability among the three methods, because its solutions tend to have a higher proportion of matrix elements that are close to zero by promoting sparsity. Through this study, we hope to have provided valuable insights into creating word embeddings with more interpretable entries. |
author2 |
Fedor Duzhin |
author_facet |
Fedor Duzhin Siow, Zi Hao |
format |
Final Year Project |
author |
Siow, Zi Hao |
author_sort |
Siow, Zi Hao |
title |
Interpretable vector language models |
title_short |
Interpretable vector language models |
title_full |
Interpretable vector language models |
title_fullStr |
Interpretable vector language models |
title_full_unstemmed |
Interpretable vector language models |
title_sort |
interpretable vector language models |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/175573 |
_version_ |
1814047364835442688 |