Interpretable vector language models

Natural Language Processing (NLP) is a branch of computer science that focuses on the development of algorithms for understanding, interpreting, and generating human language texts. A crucial technique in NLP is word embedding, where models such as Word2Vec and GloVe assign vectors to words in a voc...

Full description

Saved in:
Bibliographic Details
Main Author: Eng, Jing Keat
Other Authors: Fedor Duzhin
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2023
Subjects:
Online Access:https://hdl.handle.net/10356/166482
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-166482
record_format dspace
spelling sg-ntu-dr.10356-1664822023-05-08T15:38:35Z Interpretable vector language models Eng, Jing Keat Fedor Duzhin School of Physical and Mathematical Sciences FDuzhin@ntu.edu.sg Science::Mathematics::Algebra Science::Mathematics::Applied mathematics::Data visualization Natural Language Processing (NLP) is a branch of computer science that focuses on the development of algorithms for understanding, interpreting, and generating human language texts. A crucial technique in NLP is word embedding, where models such as Word2Vec and GloVe assign vectors to words in a vocabulary such that the Euclidean space structure (norms and angles of word vectors) aligns with the semantic structure of the training corpus. Despite their effectiveness, the individual entries of word embedding models are difficult to interpret due to the simultaneous rotation of all pre-trained word vectors preserves norms and angles while mixing up individual entries. In this study, we proposed a novel approach for generating word embeddings with interpretable entries. To achieve it, we introduced a metric to quantify the interpretability of a word embedding model. Additionally, we connected the interpretability of a word embedding model to a specific loss function defined on the Lie group SO(d). We then compared three loss functions, namely, the Varimax loss function inspired by factor analysis, the l1-norm, and a combination of the two. Our results showed that the Varimax loss function yielded word embeddings with the highest interpretability among the three methods, as it maximizes the sum of the variances of squared entries, enabling successful interpretation of some columns in the resulting word embedding matrices. This study offers insights into generating interpretable word embeddings while preserving semantic structure. Bachelor of Science in Mathematical Sciences 2023-05-02T04:50:07Z 2023-05-02T04:50:07Z 2023 Final Year Project (FYP) Eng, J. K. (2023). Interpretable vector language models. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/166482 https://hdl.handle.net/10356/166482 en application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Science::Mathematics::Algebra
Science::Mathematics::Applied mathematics::Data visualization
spellingShingle Science::Mathematics::Algebra
Science::Mathematics::Applied mathematics::Data visualization
Eng, Jing Keat
Interpretable vector language models
description Natural Language Processing (NLP) is a branch of computer science that focuses on the development of algorithms for understanding, interpreting, and generating human language texts. A crucial technique in NLP is word embedding, where models such as Word2Vec and GloVe assign vectors to words in a vocabulary such that the Euclidean space structure (norms and angles of word vectors) aligns with the semantic structure of the training corpus. Despite their effectiveness, the individual entries of word embedding models are difficult to interpret due to the simultaneous rotation of all pre-trained word vectors preserves norms and angles while mixing up individual entries. In this study, we proposed a novel approach for generating word embeddings with interpretable entries. To achieve it, we introduced a metric to quantify the interpretability of a word embedding model. Additionally, we connected the interpretability of a word embedding model to a specific loss function defined on the Lie group SO(d). We then compared three loss functions, namely, the Varimax loss function inspired by factor analysis, the l1-norm, and a combination of the two. Our results showed that the Varimax loss function yielded word embeddings with the highest interpretability among the three methods, as it maximizes the sum of the variances of squared entries, enabling successful interpretation of some columns in the resulting word embedding matrices. This study offers insights into generating interpretable word embeddings while preserving semantic structure.
author2 Fedor Duzhin
author_facet Fedor Duzhin
Eng, Jing Keat
format Final Year Project
author Eng, Jing Keat
author_sort Eng, Jing Keat
title Interpretable vector language models
title_short Interpretable vector language models
title_full Interpretable vector language models
title_fullStr Interpretable vector language models
title_full_unstemmed Interpretable vector language models
title_sort interpretable vector language models
publisher Nanyang Technological University
publishDate 2023
url https://hdl.handle.net/10356/166482
_version_ 1770563949963509760