Towards reinterpreting neural topic models via composite activations

Most Neural Topic Models (NTM) use a variational auto-encoder framework producing K topics limited to the size of the encoder’s output. These topics are interpreted through the selection of the top activated words via the weights or reconstructed vector of the decoder that are directly connected to...

Full description

Saved in:
Bibliographic Details
Main Authors: LIM, Jia Peng, LAUW, Hady Wirawan
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7610
https://ink.library.smu.edu.sg/context/sis_research/article/8613/viewcontent/emnlp22.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8613
record_format dspace
spelling sg-smu-ink.sis_research-86132022-12-22T03:28:51Z Towards reinterpreting neural topic models via composite activations LIM, Jia Peng LAUW, Hady Wirawan Most Neural Topic Models (NTM) use a variational auto-encoder framework producing K topics limited to the size of the encoder’s output. These topics are interpreted through the selection of the top activated words via the weights or reconstructed vector of the decoder that are directly connected to each neuron. In this paper, we present a model-free two-stage process to reinterpret NTM and derive further insights on the state of the trained model. Firstly, building on the original information from a trained NTM, we generate a pool of potential candidate “composite topics” by exploiting possible co-occurrences within the original set of topics, which decouples the strict interpretation of topics from the original NTM. This is followed by a combinatorial formulation to select a final set of composite topics, which we evaluate for coherence and diversity on a large external corpus. Lastly, we employ a user study to derive further insights on the reinterpretation process. 2022-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7610 https://ink.library.smu.edu.sg/context/sis_research/article/8613/viewcontent/emnlp22.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University topic modeling composite topics empirical study machine learning neural networks Artificial Intelligence and Robotics
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic topic modeling
composite topics
empirical study
machine learning
neural networks
Artificial Intelligence and Robotics
spellingShingle topic modeling
composite topics
empirical study
machine learning
neural networks
Artificial Intelligence and Robotics
LIM, Jia Peng
LAUW, Hady Wirawan
Towards reinterpreting neural topic models via composite activations
description Most Neural Topic Models (NTM) use a variational auto-encoder framework producing K topics limited to the size of the encoder’s output. These topics are interpreted through the selection of the top activated words via the weights or reconstructed vector of the decoder that are directly connected to each neuron. In this paper, we present a model-free two-stage process to reinterpret NTM and derive further insights on the state of the trained model. Firstly, building on the original information from a trained NTM, we generate a pool of potential candidate “composite topics” by exploiting possible co-occurrences within the original set of topics, which decouples the strict interpretation of topics from the original NTM. This is followed by a combinatorial formulation to select a final set of composite topics, which we evaluate for coherence and diversity on a large external corpus. Lastly, we employ a user study to derive further insights on the reinterpretation process.
format text
author LIM, Jia Peng
LAUW, Hady Wirawan
author_facet LIM, Jia Peng
LAUW, Hady Wirawan
author_sort LIM, Jia Peng
title Towards reinterpreting neural topic models via composite activations
title_short Towards reinterpreting neural topic models via composite activations
title_full Towards reinterpreting neural topic models via composite activations
title_fullStr Towards reinterpreting neural topic models via composite activations
title_full_unstemmed Towards reinterpreting neural topic models via composite activations
title_sort towards reinterpreting neural topic models via composite activations
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/7610
https://ink.library.smu.edu.sg/context/sis_research/article/8613/viewcontent/emnlp22.pdf
_version_ 1770576393912975360