Decoding the underlying meaning of multimodal hateful memes
Recent studies have proposed models that yielded promising performance for the hateful meme classification task. Nevertheless, these proposed models do not generate interpretable explanations that uncover the underlying meaning and support the classification output. A major reason for the lack of ex...
Saved in:
Main Authors: | , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2023
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8096 https://ink.library.smu.edu.sg/context/sis_research/article/9099/viewcontent/HatefulMemes_pvoa.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9099 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-90992023-09-07T07:23:26Z Decoding the underlying meaning of multimodal hateful memes HEE, Ming Shan CHONG, Wen Haw LEE, Roy Ka-Wei Recent studies have proposed models that yielded promising performance for the hateful meme classification task. Nevertheless, these proposed models do not generate interpretable explanations that uncover the underlying meaning and support the classification output. A major reason for the lack of explainable hateful meme methods is the absence of a hateful meme dataset that contains ground truth explanations for benchmarking or training. Intuitively, having such explanations can educate and assist content moderators in interpreting and removing flagged hateful memes. This paper address this research gap by introducing Hateful meme with Reasons Dataset (HatReD), which is a new multimodal hateful meme dataset annotated with the underlying hateful contextual reasons. We also define a new conditional generation task that aims to automatically generate underlying reasons to explain hateful memes and establish the baseline performance of state-of-the-art pre-trained language models on this task. We further demonstrate the usefulness of HatReD by analyzing the challenges of the new conditional generation task in explaining memes in seen and unseen domains. The dataset and benchmark models are made available here: https://github.com/Social-AI-Studio/HatRed 2023-08-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8096 info:doi/10.24963/ijcai.2023/665 https://ink.library.smu.edu.sg/context/sis_research/article/9099/viewcontent/HatefulMemes_pvoa.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Knowledge Representation and Reasoning Natural Language Processing Computer Vision Artificial Intelligence and Robotics Databases and Information Systems |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Knowledge Representation and Reasoning Natural Language Processing Computer Vision Artificial Intelligence and Robotics Databases and Information Systems |
spellingShingle |
Knowledge Representation and Reasoning Natural Language Processing Computer Vision Artificial Intelligence and Robotics Databases and Information Systems HEE, Ming Shan CHONG, Wen Haw LEE, Roy Ka-Wei Decoding the underlying meaning of multimodal hateful memes |
description |
Recent studies have proposed models that yielded promising performance for the hateful meme classification task. Nevertheless, these proposed models do not generate interpretable explanations that uncover the underlying meaning and support the classification output. A major reason for the lack of explainable hateful meme methods is the absence of a hateful meme dataset that contains ground truth explanations for benchmarking or training. Intuitively, having such explanations can educate and assist content moderators in interpreting and removing flagged hateful memes. This paper address this research gap by introducing Hateful meme with Reasons Dataset (HatReD), which is a new multimodal hateful meme dataset annotated with the underlying hateful contextual reasons. We also define a new conditional generation task that aims to automatically generate underlying reasons to explain hateful memes and establish the baseline performance of state-of-the-art pre-trained language models on this task. We further demonstrate the usefulness of HatReD by analyzing the challenges of the new conditional generation task in explaining memes in seen and unseen domains. The dataset and benchmark models are made available here: https://github.com/Social-AI-Studio/HatRed |
format |
text |
author |
HEE, Ming Shan CHONG, Wen Haw LEE, Roy Ka-Wei |
author_facet |
HEE, Ming Shan CHONG, Wen Haw LEE, Roy Ka-Wei |
author_sort |
HEE, Ming Shan |
title |
Decoding the underlying meaning of multimodal hateful memes |
title_short |
Decoding the underlying meaning of multimodal hateful memes |
title_full |
Decoding the underlying meaning of multimodal hateful memes |
title_fullStr |
Decoding the underlying meaning of multimodal hateful memes |
title_full_unstemmed |
Decoding the underlying meaning of multimodal hateful memes |
title_sort |
decoding the underlying meaning of multimodal hateful memes |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2023 |
url |
https://ink.library.smu.edu.sg/sis_research/8096 https://ink.library.smu.edu.sg/context/sis_research/article/9099/viewcontent/HatefulMemes_pvoa.pdf |
_version_ |
1779157153451343872 |