Time series task extraction from large language models
Recent advancements in large language models (LLMs) have shown tremendous potential to revolutionize time series classification. These models possess newly improved capabilities, including impressive zero-shot learning and remarkable reasoning skills, without requiring any additional training...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/180995 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-180995 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1809952024-11-11T01:21:17Z Time series task extraction from large language models Toh, Leong Seng Thomas Peyrin College of Computing and Data Science thomas.peyrin@ntu.edu.sg Computer and Information Science Large language models (LLMs) Time-series classification Zero-shot learning Rule extraction Recent advancements in large language models (LLMs) have shown tremendous potential to revolutionize time series classification. These models possess newly improved capabilities, including impressive zero-shot learning and remarkable reasoning skills, without requiring any additional training data. We anticipate that ALLM will become the standard for time series classification, eventually replacing resource-intensive machine learning models. However, the lack of interpretability in LLMsandtheir potential for inaccuracies pose significant challenges that undermine user trust. To build user trust, two critical gaps need to be addressed: reliability and interpretability. To address this issue, we propose a method to approximate ALLM using human-interpretable binary feature rules, denoted as ¯ Arule. This approach leverages the TT-rules (Truth Table rules) model developed by Benamira et al., 2023 to extract binary rules through LLM inference on time series datasets. The LLM is set aside once the rules are derived and inference is conducted exclusively using Arule. This methodology will be validated using three cyber-security datasets, while incorporating the privacy-preserving features outlined by Soegeng, 2024 to ensure the protection of sensitive data. Bachelor's degree 2024-11-11T01:21:17Z 2024-11-11T01:21:17Z 2024 Final Year Project (FYP) Toh, L. S. (2024). Time series task extraction from large language models. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/180995 https://hdl.handle.net/10356/180995 en SCSE23-1020 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Large language models (LLMs) Time-series classification Zero-shot learning Rule extraction |
spellingShingle |
Computer and Information Science Large language models (LLMs) Time-series classification Zero-shot learning Rule extraction Toh, Leong Seng Time series task extraction from large language models |
description |
Recent advancements in large language models (LLMs) have shown tremendous potential
to revolutionize time series classification. These models possess newly improved capabilities,
including impressive zero-shot learning and remarkable reasoning skills, without requiring any
additional training data.
We anticipate that ALLM will become the standard for time series classification, eventually
replacing resource-intensive machine learning models. However, the lack of interpretability in
LLMsandtheir potential for inaccuracies pose significant challenges that undermine user trust.
To build user trust, two critical gaps need to be addressed: reliability and interpretability.
To address this issue, we propose a method to approximate ALLM using human-interpretable
binary feature rules, denoted as ¯ Arule. This approach leverages the TT-rules (Truth Table rules)
model developed by Benamira et al., 2023 to extract binary rules through LLM inference on
time series datasets. The LLM is set aside once the rules are derived and inference is conducted
exclusively using Arule. This methodology will be validated using three cyber-security datasets,
while incorporating the privacy-preserving features outlined by Soegeng, 2024 to ensure the
protection of sensitive data. |
author2 |
Thomas Peyrin |
author_facet |
Thomas Peyrin Toh, Leong Seng |
format |
Final Year Project |
author |
Toh, Leong Seng |
author_sort |
Toh, Leong Seng |
title |
Time series task extraction from large language models |
title_short |
Time series task extraction from large language models |
title_full |
Time series task extraction from large language models |
title_fullStr |
Time series task extraction from large language models |
title_full_unstemmed |
Time series task extraction from large language models |
title_sort |
time series task extraction from large language models |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/180995 |
_version_ |
1816859047662452736 |