Towards reinterpreting neural topic models via composite activations
Most Neural Topic Models (NTM) use a variational auto-encoder framework producing K topics limited to the size of the encoder’s output. These topics are interpreted through the selection of the top activated words via the weights or reconstructed vector of the decoder that are directly connected to...
Saved in:
Main Authors: | , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2022
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/7610 https://ink.library.smu.edu.sg/context/sis_research/article/8613/viewcontent/emnlp22.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-8613 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-86132022-12-22T03:28:51Z Towards reinterpreting neural topic models via composite activations LIM, Jia Peng LAUW, Hady Wirawan Most Neural Topic Models (NTM) use a variational auto-encoder framework producing K topics limited to the size of the encoder’s output. These topics are interpreted through the selection of the top activated words via the weights or reconstructed vector of the decoder that are directly connected to each neuron. In this paper, we present a model-free two-stage process to reinterpret NTM and derive further insights on the state of the trained model. Firstly, building on the original information from a trained NTM, we generate a pool of potential candidate “composite topics” by exploiting possible co-occurrences within the original set of topics, which decouples the strict interpretation of topics from the original NTM. This is followed by a combinatorial formulation to select a final set of composite topics, which we evaluate for coherence and diversity on a large external corpus. Lastly, we employ a user study to derive further insights on the reinterpretation process. 2022-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7610 https://ink.library.smu.edu.sg/context/sis_research/article/8613/viewcontent/emnlp22.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University topic modeling composite topics empirical study machine learning neural networks Artificial Intelligence and Robotics |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
topic modeling composite topics empirical study machine learning neural networks Artificial Intelligence and Robotics |
spellingShingle |
topic modeling composite topics empirical study machine learning neural networks Artificial Intelligence and Robotics LIM, Jia Peng LAUW, Hady Wirawan Towards reinterpreting neural topic models via composite activations |
description |
Most Neural Topic Models (NTM) use a variational auto-encoder framework producing K topics limited to the size of the encoder’s output. These topics are interpreted through the selection of the top activated words via the weights or reconstructed vector of the decoder that are directly connected to each neuron. In this paper, we present a model-free two-stage process to reinterpret NTM and derive further insights on the state of the trained model. Firstly, building on the original information from a trained NTM, we generate a pool of potential candidate “composite topics” by exploiting possible co-occurrences within the original set of topics, which decouples the strict interpretation of topics from the original NTM. This is followed by a combinatorial formulation to select a final set of composite topics, which we evaluate for coherence and diversity on a large external corpus. Lastly, we employ a user study to derive further insights on the reinterpretation process. |
format |
text |
author |
LIM, Jia Peng LAUW, Hady Wirawan |
author_facet |
LIM, Jia Peng LAUW, Hady Wirawan |
author_sort |
LIM, Jia Peng |
title |
Towards reinterpreting neural topic models via composite activations |
title_short |
Towards reinterpreting neural topic models via composite activations |
title_full |
Towards reinterpreting neural topic models via composite activations |
title_fullStr |
Towards reinterpreting neural topic models via composite activations |
title_full_unstemmed |
Towards reinterpreting neural topic models via composite activations |
title_sort |
towards reinterpreting neural topic models via composite activations |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2022 |
url |
https://ink.library.smu.edu.sg/sis_research/7610 https://ink.library.smu.edu.sg/context/sis_research/article/8613/viewcontent/emnlp22.pdf |
_version_ |
1770576393912975360 |