Framework to evaluate and test defences against hallucination in large language model
The recent advancement of AI, particularly the large language models (LLMs) has en- abled unprecedented capabilities in natural language processing (NLP) tasks, including things such as content generation, translation, and question answering (QA). However, just like any new technology, LLMs faced...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/180892 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-180892 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1808922024-11-04T01:10:23Z Framework to evaluate and test defences against hallucination in large language model Pan, Johnny Shi Han Luu Anh Tuan College of Computing and Data Science anhtuan.luu@ntu.edu.sg Computer and Information Science Large language model LLM safety LLM testing and defences The recent advancement of AI, particularly the large language models (LLMs) has en- abled unprecedented capabilities in natural language processing (NLP) tasks, including things such as content generation, translation, and question answering (QA). However, just like any new technology, LLMs faced some challenges. One of the key issues with LLMs is what’s known as “hallucination.” This happens when the model produces information that is incorrect or made up but still sounds plausible. In this paper, the goal is to outline a framework to help identify and assess these hallucinations through curating new hallucination evaluation methods, datasets and evaluation metrics. The framework is adaptable and can be used with a variety of models and LLMs prod- uct. The main objective is to offer developers and engineers a consistent approach to identifying hallucinations in LLM-based applications before they are released. Bachelor's degree 2024-11-04T01:10:23Z 2024-11-04T01:10:23Z 2024 Final Year Project (FYP) Pan, J. S. H. (2024). Framework to evaluate and test defences against hallucination in large language model. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/180892 https://hdl.handle.net/10356/180892 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Large language model LLM safety LLM testing and defences |
spellingShingle |
Computer and Information Science Large language model LLM safety LLM testing and defences Pan, Johnny Shi Han Framework to evaluate and test defences against hallucination in large language model |
description |
The recent advancement of AI, particularly the large language models (LLMs) has en-
abled unprecedented capabilities in natural language processing (NLP) tasks, including
things such as content generation, translation, and question answering (QA). However,
just like any new technology, LLMs faced some challenges. One of the key issues
with LLMs is what’s known as “hallucination.” This happens when the model produces
information that is incorrect or made up but still sounds plausible. In this paper, the
goal is to outline a framework to help identify and assess these hallucinations through
curating new hallucination evaluation methods, datasets and evaluation metrics. The
framework is adaptable and can be used with a variety of models and LLMs prod-
uct. The main objective is to offer developers and engineers a consistent approach to
identifying hallucinations in LLM-based applications before they are released. |
author2 |
Luu Anh Tuan |
author_facet |
Luu Anh Tuan Pan, Johnny Shi Han |
format |
Final Year Project |
author |
Pan, Johnny Shi Han |
author_sort |
Pan, Johnny Shi Han |
title |
Framework to evaluate and test defences against hallucination in large language model |
title_short |
Framework to evaluate and test defences against hallucination in large language model |
title_full |
Framework to evaluate and test defences against hallucination in large language model |
title_fullStr |
Framework to evaluate and test defences against hallucination in large language model |
title_full_unstemmed |
Framework to evaluate and test defences against hallucination in large language model |
title_sort |
framework to evaluate and test defences against hallucination in large language model |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/180892 |
_version_ |
1816859006574002176 |