MMEKG: Multi-modal Event Knowledge Graph towards universal representation across modalities

Events are fundamental building blocks of realworld happenings. In this paper, we present a large-scale, multi-modal event knowledge graph named MMEKG. MMEKG unifies different modalities of knowledge via events, which complement and disambiguate each other. Specifically, MMEKG incorporates (i) over...

Full description

Saved in:
Bibliographic Details
Main Authors: MA, Yubo, WANG, Zehao, LI, Mukai, CAO, Yixin, CHEN, Meiqi, LI, Xinze, SUN, Wenqi, DENG, Kunquan, WANG, Kun, SUN, Aixin, SHAO, Jing
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7445
https://ink.library.smu.edu.sg/context/sis_research/article/8448/viewcontent/2022.acl_demo.23.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8448
record_format dspace
spelling sg-smu-ink.sis_research-84482022-10-20T07:37:35Z MMEKG: Multi-modal Event Knowledge Graph towards universal representation across modalities MA, Yubo WANG, Zehao LI, Mukai CAO, Yixin CHEN, Meiqi LI, Xinze SUN, Wenqi DENG, Kunquan WANG, Kun SUN, Aixin SHAO, Jing Events are fundamental building blocks of realworld happenings. In this paper, we present a large-scale, multi-modal event knowledge graph named MMEKG. MMEKG unifies different modalities of knowledge via events, which complement and disambiguate each other. Specifically, MMEKG incorporates (i) over 990 thousand concept events with 644 relation types to cover most types of happenings, and (ii) over 863 million instance events connected through 934 million relations, which provide rich contextual information in texts and/or images. To collect billion-scale instance events and relations among them, we additionally develop an efficient yet effective pipeline for textual/visual knowledge extraction system. We also develop an induction strategy to create million-scale concept events and a schema organizing all events and relations in MMEKG. To this end, we also provide a pipeline1 enabling our system to seamlessly parse texts/images to event graphs and to retrieve multi-modal knowledge at both concept- and instance-levels. 2022-05-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7445 info:doi/10.18653/v1/2022.acl-demo.23 https://ink.library.smu.edu.sg/context/sis_research/article/8448/viewcontent/2022.acl_demo.23.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems Graphics and Human Computer Interfaces
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Databases and Information Systems
Graphics and Human Computer Interfaces
spellingShingle Databases and Information Systems
Graphics and Human Computer Interfaces
MA, Yubo
WANG, Zehao
LI, Mukai
CAO, Yixin
CHEN, Meiqi
LI, Xinze
SUN, Wenqi
DENG, Kunquan
WANG, Kun
SUN, Aixin
SHAO, Jing
MMEKG: Multi-modal Event Knowledge Graph towards universal representation across modalities
description Events are fundamental building blocks of realworld happenings. In this paper, we present a large-scale, multi-modal event knowledge graph named MMEKG. MMEKG unifies different modalities of knowledge via events, which complement and disambiguate each other. Specifically, MMEKG incorporates (i) over 990 thousand concept events with 644 relation types to cover most types of happenings, and (ii) over 863 million instance events connected through 934 million relations, which provide rich contextual information in texts and/or images. To collect billion-scale instance events and relations among them, we additionally develop an efficient yet effective pipeline for textual/visual knowledge extraction system. We also develop an induction strategy to create million-scale concept events and a schema organizing all events and relations in MMEKG. To this end, we also provide a pipeline1 enabling our system to seamlessly parse texts/images to event graphs and to retrieve multi-modal knowledge at both concept- and instance-levels.
format text
author MA, Yubo
WANG, Zehao
LI, Mukai
CAO, Yixin
CHEN, Meiqi
LI, Xinze
SUN, Wenqi
DENG, Kunquan
WANG, Kun
SUN, Aixin
SHAO, Jing
author_facet MA, Yubo
WANG, Zehao
LI, Mukai
CAO, Yixin
CHEN, Meiqi
LI, Xinze
SUN, Wenqi
DENG, Kunquan
WANG, Kun
SUN, Aixin
SHAO, Jing
author_sort MA, Yubo
title MMEKG: Multi-modal Event Knowledge Graph towards universal representation across modalities
title_short MMEKG: Multi-modal Event Knowledge Graph towards universal representation across modalities
title_full MMEKG: Multi-modal Event Knowledge Graph towards universal representation across modalities
title_fullStr MMEKG: Multi-modal Event Knowledge Graph towards universal representation across modalities
title_full_unstemmed MMEKG: Multi-modal Event Knowledge Graph towards universal representation across modalities
title_sort mmekg: multi-modal event knowledge graph towards universal representation across modalities
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/7445
https://ink.library.smu.edu.sg/context/sis_research/article/8448/viewcontent/2022.acl_demo.23.pdf
_version_ 1770576350427480064