Image enhanced event detection in news articles
Event detection is a crucial and challenging sub-task of event extraction, which suffers from a severe ambiguity issue of trigger words. Existing works mainly focus on using textual context information, while there naturally exist many images accompanied by news articles that are yet to be explored....
Saved in:
Main Authors: | , , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2020
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/7456 https://ink.library.smu.edu.sg/context/sis_research/article/8459/viewcontent/6437_Article_Text_9662_1_10_20200517.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-8459 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-84592022-10-20T07:19:27Z Image enhanced event detection in news articles TONG, Meihan WANG, Shuai CAO, Yixin XU, Bin LI, Juaizi HOU, Lei CHUA, Tat-Seng Event detection is a crucial and challenging sub-task of event extraction, which suffers from a severe ambiguity issue of trigger words. Existing works mainly focus on using textual context information, while there naturally exist many images accompanied by news articles that are yet to be explored. We believe that images not only reflect the core events of the text, but are also helpful for the disambiguation of trigger words. In this paper, we first contribute an image dataset supplement to ED benchmarks (i.e., ACE2005) for training and evaluation. We then propose a novel Dual Recurrent Multimodal Model, DRMM, to conduct deep interactions between images and sentences for modality features aggregation. DRMM utilizes pre-trained BERT and ResNet to encode sentences and images, and employs an alternating dual attention to select informative features for mutual enhancements. Our superior performance compared to six state-of-art baselines as well as further ablation studies demonstrate the significance of image modality and effectiveness of the proposed architecture. The code and image dataset are avaliable at https://github.com/ shuaiwa16/image-enhanced-event-extraction. 2020-02-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7456 info:doi/10.1609/aaai.v34i05.6437 https://ink.library.smu.edu.sg/context/sis_research/article/8459/viewcontent/6437_Article_Text_9662_1_10_20200517.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems Graphics and Human Computer Interfaces |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Databases and Information Systems Graphics and Human Computer Interfaces |
spellingShingle |
Databases and Information Systems Graphics and Human Computer Interfaces TONG, Meihan WANG, Shuai CAO, Yixin XU, Bin LI, Juaizi HOU, Lei CHUA, Tat-Seng Image enhanced event detection in news articles |
description |
Event detection is a crucial and challenging sub-task of event extraction, which suffers from a severe ambiguity issue of trigger words. Existing works mainly focus on using textual context information, while there naturally exist many images accompanied by news articles that are yet to be explored. We believe that images not only reflect the core events of the text, but are also helpful for the disambiguation of trigger words. In this paper, we first contribute an image dataset supplement to ED benchmarks (i.e., ACE2005) for training and evaluation. We then propose a novel Dual Recurrent Multimodal Model, DRMM, to conduct deep interactions between images and sentences for modality features aggregation. DRMM utilizes pre-trained BERT and ResNet to encode sentences and images, and employs an alternating dual attention to select informative features for mutual enhancements. Our superior performance compared to six state-of-art baselines as well as further ablation studies demonstrate the significance of image modality and effectiveness of the proposed architecture. The code and image dataset are avaliable at https://github.com/ shuaiwa16/image-enhanced-event-extraction. |
format |
text |
author |
TONG, Meihan WANG, Shuai CAO, Yixin XU, Bin LI, Juaizi HOU, Lei CHUA, Tat-Seng |
author_facet |
TONG, Meihan WANG, Shuai CAO, Yixin XU, Bin LI, Juaizi HOU, Lei CHUA, Tat-Seng |
author_sort |
TONG, Meihan |
title |
Image enhanced event detection in news articles |
title_short |
Image enhanced event detection in news articles |
title_full |
Image enhanced event detection in news articles |
title_fullStr |
Image enhanced event detection in news articles |
title_full_unstemmed |
Image enhanced event detection in news articles |
title_sort |
image enhanced event detection in news articles |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2020 |
url |
https://ink.library.smu.edu.sg/sis_research/7456 https://ink.library.smu.edu.sg/context/sis_research/article/8459/viewcontent/6437_Article_Text_9662_1_10_20200517.pdf |
_version_ |
1770576341777776640 |