Analyzing temporal complex events with large language models? A benchmark towards temporal, long context understanding
The digital landscape is rapidly evolving with an ever-increasing volume of online news, emphasizing the need for swift and precise analysis of complex events.We refer to the complex events composed of many news articles over an extended period as Temporal Complex Event (TCE). This paper proposes a...
Saved in:
Main Authors: | , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2024
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/9697 https://ink.library.smu.edu.sg/context/sis_research/article/10697/viewcontent/2024.acl_long.87.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-10697 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-106972024-11-28T09:04:10Z Analyzing temporal complex events with large language models? A benchmark towards temporal, long context understanding ZHANG, Zhihan CAO, Yixin YE, Chenchen MA. Yunshan, LIAO, Lizi CHUA, Tat-Seng The digital landscape is rapidly evolving with an ever-increasing volume of online news, emphasizing the need for swift and precise analysis of complex events.We refer to the complex events composed of many news articles over an extended period as Temporal Complex Event (TCE). This paper proposes a novel approach using Large Language Models (LLMs) to systematically extract and analyze the event chain within TCE, characterized by their key points and timestamps. We establish a benchmark, named TCELongBench, to evaluate the proficiency of LLMs in handling temporal dynamics and understanding extensive text. This benchmark encompasses three distinct tasks - reading comprehension, temporal sequencing, and future event forecasting. In the experiment, we leverage retrieval-augmented generation (RAG) method and LLMs with long context window to deal with lengthy news articles of TCE. Our findings indicate that models with suitable retrievers exhibit comparable performance with those utilizing long context window. 2024-08-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9697 info:doi/10.18653/v1/2024.acl-long.87 https://ink.library.smu.edu.sg/context/sis_research/article/10697/viewcontent/2024.acl_long.87.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Temporal complex events Large language models LLMS Extensive text processing Artificial Intelligence and Robotics Computer Sciences |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Temporal complex events Large language models LLMS Extensive text processing Artificial Intelligence and Robotics Computer Sciences |
spellingShingle |
Temporal complex events Large language models LLMS Extensive text processing Artificial Intelligence and Robotics Computer Sciences ZHANG, Zhihan CAO, Yixin YE, Chenchen MA. Yunshan, LIAO, Lizi CHUA, Tat-Seng Analyzing temporal complex events with large language models? A benchmark towards temporal, long context understanding |
description |
The digital landscape is rapidly evolving with an ever-increasing volume of online news, emphasizing the need for swift and precise analysis of complex events.We refer to the complex events composed of many news articles over an extended period as Temporal Complex Event (TCE). This paper proposes a novel approach using Large Language Models (LLMs) to systematically extract and analyze the event chain within TCE, characterized by their key points and timestamps. We establish a benchmark, named TCELongBench, to evaluate the proficiency of LLMs in handling temporal dynamics and understanding extensive text. This benchmark encompasses three distinct tasks - reading comprehension, temporal sequencing, and future event forecasting. In the experiment, we leverage retrieval-augmented generation (RAG) method and LLMs with long context window to deal with lengthy news articles of TCE. Our findings indicate that models with suitable retrievers exhibit comparable performance with those utilizing long context window. |
format |
text |
author |
ZHANG, Zhihan CAO, Yixin YE, Chenchen MA. Yunshan, LIAO, Lizi CHUA, Tat-Seng |
author_facet |
ZHANG, Zhihan CAO, Yixin YE, Chenchen MA. Yunshan, LIAO, Lizi CHUA, Tat-Seng |
author_sort |
ZHANG, Zhihan |
title |
Analyzing temporal complex events with large language models? A benchmark towards temporal, long context understanding |
title_short |
Analyzing temporal complex events with large language models? A benchmark towards temporal, long context understanding |
title_full |
Analyzing temporal complex events with large language models? A benchmark towards temporal, long context understanding |
title_fullStr |
Analyzing temporal complex events with large language models? A benchmark towards temporal, long context understanding |
title_full_unstemmed |
Analyzing temporal complex events with large language models? A benchmark towards temporal, long context understanding |
title_sort |
analyzing temporal complex events with large language models? a benchmark towards temporal, long context understanding |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2024 |
url |
https://ink.library.smu.edu.sg/sis_research/9697 https://ink.library.smu.edu.sg/context/sis_research/article/10697/viewcontent/2024.acl_long.87.pdf |
_version_ |
1819113105920098304 |