Clustering word embeddings with different properties for topic modelling
The goal of topic detection or topic modelling is to uncover the hidden topics in a large corpus. It’s an increasingly useful analysis tool in the information age. As research progresses, topic modelling methods have gradually expanded from probabilistic methods to distributed representations. The e...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Master by Coursework |
Language: | English |
Published: |
Nanyang Technological University
2021
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/152557 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-152557 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1525572023-07-04T17:35:49Z Clustering word embeddings with different properties for topic modelling Wu, Yijun Lihui Chen School of Electrical and Electronic Engineering ELHCHEN@ntu.edu.sg Engineering::Electrical and electronic engineering The goal of topic detection or topic modelling is to uncover the hidden topics in a large corpus. It’s an increasingly useful analysis tool in the information age. As research progresses, topic modelling methods have gradually expanded from probabilistic methods to distributed representations. The emergence of BERT announced a huge change in the NLP paradigm. This deep model pre-trained from the unlabeled datasets significantly improved the accuracy of the NLP task, and the topic modelling method was further extended to use pre-trained word embeddings to complete. The success has inspired more BERT-based models. This project is about topic modelling with machine learning techniques. Pre-trained or fine-tuned embeddings from BERT, SBERT, ERNIE 2.0 or SimCSE are applied together with weighted K-means, where document statistics are used as weighting factor, to group or re-rank top words to identify top-20 topics using open source 20 newsgroups dataset. The results show that the SBERT and SimCSE_sup models outperform the others. In addition, the properties of different models embeddings and their impact on topic identification have also been discussed. Master of Science (Communications Engineering) 2021-08-31T05:35:28Z 2021-08-31T05:35:28Z 2021 Thesis-Master by Coursework Wu, Y. (2021). Clustering word embeddings with different properties for topic modelling. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/152557 https://hdl.handle.net/10356/152557 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Electrical and electronic engineering |
spellingShingle |
Engineering::Electrical and electronic engineering Wu, Yijun Clustering word embeddings with different properties for topic modelling |
description |
The goal of topic detection or topic modelling is to uncover the hidden topics in a large corpus. It’s an increasingly useful analysis tool in the information age. As research progresses, topic modelling methods have gradually expanded from probabilistic methods to distributed representations. The emergence of BERT announced a huge change in the NLP paradigm. This deep model pre-trained from the unlabeled datasets significantly improved the accuracy of the NLP task, and the topic modelling method was further extended to use pre-trained word embeddings to complete. The success has inspired more BERT-based models. This project is about topic modelling with machine learning techniques. Pre-trained or fine-tuned embeddings from BERT, SBERT, ERNIE 2.0 or SimCSE are applied together with weighted K-means, where document statistics are used as weighting factor, to group or re-rank top words to identify top-20 topics using open source 20 newsgroups dataset. The results show that the SBERT and SimCSE_sup models outperform the others. In addition, the properties of different models embeddings and their impact on topic identification have also been discussed. |
author2 |
Lihui Chen |
author_facet |
Lihui Chen Wu, Yijun |
format |
Thesis-Master by Coursework |
author |
Wu, Yijun |
author_sort |
Wu, Yijun |
title |
Clustering word embeddings with different properties for topic modelling |
title_short |
Clustering word embeddings with different properties for topic modelling |
title_full |
Clustering word embeddings with different properties for topic modelling |
title_fullStr |
Clustering word embeddings with different properties for topic modelling |
title_full_unstemmed |
Clustering word embeddings with different properties for topic modelling |
title_sort |
clustering word embeddings with different properties for topic modelling |
publisher |
Nanyang Technological University |
publishDate |
2021 |
url |
https://hdl.handle.net/10356/152557 |
_version_ |
1772828233916481536 |