Interpretable vector language models

Natural Language Processing (NLP) is an important part of Artificial Intelligence (AI) that aims to create algorithms which improve how humans understand and interpret bodies of text. In particular, word embeddings form a vital part of NLP, as models like Word2Vec and GloVe assign numeric vectors to...

Full description

Saved in:
Bibliographic Details
Main Author: Siow, Zi Hao
Other Authors: Fedor Duzhin
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/175573
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-175573
record_format dspace
spelling sg-ntu-dr.10356-1755732024-05-06T15:37:34Z Interpretable vector language models Siow, Zi Hao Fedor Duzhin School of Physical and Mathematical Sciences FDuzhin@ntu.edu.sg Mathematical Sciences Interpretable Vector language models Data visualisation Natural Language Processing (NLP) is an important part of Artificial Intelligence (AI) that aims to create algorithms which improve how humans understand and interpret bodies of text. In particular, word embeddings form a vital part of NLP, as models like Word2Vec and GloVe assign numeric vectors to words in a text corpus such that norms and angles between words are preserved and semantic structure is maintained. While their effectiveness is undisputed, they face a major limitation in the form of limited interpretability, as individual entries are hard to interpret due to the simultaneous rotation of all vectors preserving semantic structure while entries become mixed up. Hence, in this study, we proposed a novel approach of generating word embeddings with a higher degree of interpretability. We associated the interpretability of a word embedding with the optimisation of various loss functions, namely Varimax, Quartimax and the l1-norm, defined on the Lie group SO(d). Our findings revealed that the l1-norm method achieved the highest level of interpretability among the three methods, because its solutions tend to have a higher proportion of matrix elements that are close to zero by promoting sparsity. Through this study, we hope to have provided valuable insights into creating word embeddings with more interpretable entries. Bachelor's degree 2024-04-30T03:57:24Z 2024-04-30T03:57:24Z 2024 Final Year Project (FYP) Siow, Z. H. (2024). Interpretable vector language models. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/175573 https://hdl.handle.net/10356/175573 en application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Mathematical Sciences
Interpretable
Vector language models
Data visualisation
spellingShingle Mathematical Sciences
Interpretable
Vector language models
Data visualisation
Siow, Zi Hao
Interpretable vector language models
description Natural Language Processing (NLP) is an important part of Artificial Intelligence (AI) that aims to create algorithms which improve how humans understand and interpret bodies of text. In particular, word embeddings form a vital part of NLP, as models like Word2Vec and GloVe assign numeric vectors to words in a text corpus such that norms and angles between words are preserved and semantic structure is maintained. While their effectiveness is undisputed, they face a major limitation in the form of limited interpretability, as individual entries are hard to interpret due to the simultaneous rotation of all vectors preserving semantic structure while entries become mixed up. Hence, in this study, we proposed a novel approach of generating word embeddings with a higher degree of interpretability. We associated the interpretability of a word embedding with the optimisation of various loss functions, namely Varimax, Quartimax and the l1-norm, defined on the Lie group SO(d). Our findings revealed that the l1-norm method achieved the highest level of interpretability among the three methods, because its solutions tend to have a higher proportion of matrix elements that are close to zero by promoting sparsity. Through this study, we hope to have provided valuable insights into creating word embeddings with more interpretable entries.
author2 Fedor Duzhin
author_facet Fedor Duzhin
Siow, Zi Hao
format Final Year Project
author Siow, Zi Hao
author_sort Siow, Zi Hao
title Interpretable vector language models
title_short Interpretable vector language models
title_full Interpretable vector language models
title_fullStr Interpretable vector language models
title_full_unstemmed Interpretable vector language models
title_sort interpretable vector language models
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/175573
_version_ 1814047364835442688