Information screening whilst exploiting! multimodal relation extraction with feature denoising and multimodal topic modeling
Existing research on multimodal relation extraction (MRE) faces two co-existing challenges, internal-information over-utilization and external-information under-exploitation. To combat that, we propose a novel framework that simultaneously implements the idea of internal-information screening and ex...
Saved in:
Main Authors: | , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2023
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8261 https://ink.library.smu.edu.sg/context/sis_research/article/9264/viewcontent/2023.acl_long.823__1_.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9264 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-92642023-11-10T08:55:54Z Information screening whilst exploiting! multimodal relation extraction with feature denoising and multimodal topic modeling WU, Shengqiong FEI, Hao CAO, Yixin BING, Lidong CHUA, Tat-Seng Existing research on multimodal relation extraction (MRE) faces two co-existing challenges, internal-information over-utilization and external-information under-exploitation. To combat that, we propose a novel framework that simultaneously implements the idea of internal-information screening and external-information exploiting. First, we represent the fine-grained semantic structures of the input image and text with the visual and textual scene graphs, which are further fused into a unified cross-modal graph (CMG). Based on CMG, we perform structure refinement with the guidance of the graph information bottleneck principle, actively denoising the less-informative features. Next, we perform topic modeling over the input image and text, incorporating latent multimodal topic features to enrich the contexts. On the benchmark MRE dataset, our system outperforms the current best model significantly. With further in-depth analyses, we reveal the great potential of our method for the MRE task. 2023-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8261 info:doi/10.18653/v1/2023.acl-long.823 https://ink.library.smu.edu.sg/context/sis_research/article/9264/viewcontent/2023.acl_long.823__1_.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University computational linguistics; data mining; extraction; information retrieval Computer Sciences |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
computational linguistics; data mining; extraction; information retrieval Computer Sciences |
spellingShingle |
computational linguistics; data mining; extraction; information retrieval Computer Sciences WU, Shengqiong FEI, Hao CAO, Yixin BING, Lidong CHUA, Tat-Seng Information screening whilst exploiting! multimodal relation extraction with feature denoising and multimodal topic modeling |
description |
Existing research on multimodal relation extraction (MRE) faces two co-existing challenges, internal-information over-utilization and external-information under-exploitation. To combat that, we propose a novel framework that simultaneously implements the idea of internal-information screening and external-information exploiting. First, we represent the fine-grained semantic structures of the input image and text with the visual and textual scene graphs, which are further fused into a unified cross-modal graph (CMG). Based on CMG, we perform structure refinement with the guidance of the graph information bottleneck principle, actively denoising the less-informative features. Next, we perform topic modeling over the input image and text, incorporating latent multimodal topic features to enrich the contexts. On the benchmark MRE dataset, our system outperforms the current best model significantly. With further in-depth analyses, we reveal the great potential of our method for the MRE task. |
format |
text |
author |
WU, Shengqiong FEI, Hao CAO, Yixin BING, Lidong CHUA, Tat-Seng |
author_facet |
WU, Shengqiong FEI, Hao CAO, Yixin BING, Lidong CHUA, Tat-Seng |
author_sort |
WU, Shengqiong |
title |
Information screening whilst exploiting! multimodal relation extraction with feature denoising and multimodal topic modeling |
title_short |
Information screening whilst exploiting! multimodal relation extraction with feature denoising and multimodal topic modeling |
title_full |
Information screening whilst exploiting! multimodal relation extraction with feature denoising and multimodal topic modeling |
title_fullStr |
Information screening whilst exploiting! multimodal relation extraction with feature denoising and multimodal topic modeling |
title_full_unstemmed |
Information screening whilst exploiting! multimodal relation extraction with feature denoising and multimodal topic modeling |
title_sort |
information screening whilst exploiting! multimodal relation extraction with feature denoising and multimodal topic modeling |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2023 |
url |
https://ink.library.smu.edu.sg/sis_research/8261 https://ink.library.smu.edu.sg/context/sis_research/article/9264/viewcontent/2023.acl_long.823__1_.pdf |
_version_ |
1783955659805425664 |