Time series task extraction from large language models

Recent advancements in large language models (LLMs) have shown tremendous potential to revolutionize time series classification. These models possess newly improved capabilities, including impressive zero-shot learning and remarkable reasoning skills, without requiring any additional training...

Full description

Saved in:
Bibliographic Details
Main Author: Toh, Leong Seng
Other Authors: Thomas Peyrin
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/180995
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-180995
record_format dspace
spelling sg-ntu-dr.10356-1809952024-11-11T01:21:17Z Time series task extraction from large language models Toh, Leong Seng Thomas Peyrin College of Computing and Data Science thomas.peyrin@ntu.edu.sg Computer and Information Science Large language models (LLMs) Time-series classification Zero-shot learning Rule extraction Recent advancements in large language models (LLMs) have shown tremendous potential to revolutionize time series classification. These models possess newly improved capabilities, including impressive zero-shot learning and remarkable reasoning skills, without requiring any additional training data. We anticipate that ALLM will become the standard for time series classification, eventually replacing resource-intensive machine learning models. However, the lack of interpretability in LLMsandtheir potential for inaccuracies pose significant challenges that undermine user trust. To build user trust, two critical gaps need to be addressed: reliability and interpretability. To address this issue, we propose a method to approximate ALLM using human-interpretable binary feature rules, denoted as ¯ Arule. This approach leverages the TT-rules (Truth Table rules) model developed by Benamira et al., 2023 to extract binary rules through LLM inference on time series datasets. The LLM is set aside once the rules are derived and inference is conducted exclusively using Arule. This methodology will be validated using three cyber-security datasets, while incorporating the privacy-preserving features outlined by Soegeng, 2024 to ensure the protection of sensitive data. Bachelor's degree 2024-11-11T01:21:17Z 2024-11-11T01:21:17Z 2024 Final Year Project (FYP) Toh, L. S. (2024). Time series task extraction from large language models. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/180995 https://hdl.handle.net/10356/180995 en SCSE23-1020 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
Large language models (LLMs)
Time-series classification
Zero-shot learning
Rule extraction
spellingShingle Computer and Information Science
Large language models (LLMs)
Time-series classification
Zero-shot learning
Rule extraction
Toh, Leong Seng
Time series task extraction from large language models
description Recent advancements in large language models (LLMs) have shown tremendous potential to revolutionize time series classification. These models possess newly improved capabilities, including impressive zero-shot learning and remarkable reasoning skills, without requiring any additional training data. We anticipate that ALLM will become the standard for time series classification, eventually replacing resource-intensive machine learning models. However, the lack of interpretability in LLMsandtheir potential for inaccuracies pose significant challenges that undermine user trust. To build user trust, two critical gaps need to be addressed: reliability and interpretability. To address this issue, we propose a method to approximate ALLM using human-interpretable binary feature rules, denoted as ¯ Arule. This approach leverages the TT-rules (Truth Table rules) model developed by Benamira et al., 2023 to extract binary rules through LLM inference on time series datasets. The LLM is set aside once the rules are derived and inference is conducted exclusively using Arule. This methodology will be validated using three cyber-security datasets, while incorporating the privacy-preserving features outlined by Soegeng, 2024 to ensure the protection of sensitive data.
author2 Thomas Peyrin
author_facet Thomas Peyrin
Toh, Leong Seng
format Final Year Project
author Toh, Leong Seng
author_sort Toh, Leong Seng
title Time series task extraction from large language models
title_short Time series task extraction from large language models
title_full Time series task extraction from large language models
title_fullStr Time series task extraction from large language models
title_full_unstemmed Time series task extraction from large language models
title_sort time series task extraction from large language models
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/180995
_version_ 1816859047662452736