Framework to evaluate and test defences against hallucination in large language model

The recent advancement of AI, particularly the large language models (LLMs) has en- abled unprecedented capabilities in natural language processing (NLP) tasks, including things such as content generation, translation, and question answering (QA). However, just like any new technology, LLMs faced...

Full description

Saved in:
Bibliographic Details
Main Author: Pan, Johnny Shi Han
Other Authors: Luu Anh Tuan
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/180892
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-180892
record_format dspace
spelling sg-ntu-dr.10356-1808922024-11-04T01:10:23Z Framework to evaluate and test defences against hallucination in large language model Pan, Johnny Shi Han Luu Anh Tuan College of Computing and Data Science anhtuan.luu@ntu.edu.sg Computer and Information Science Large language model LLM safety LLM testing and defences The recent advancement of AI, particularly the large language models (LLMs) has en- abled unprecedented capabilities in natural language processing (NLP) tasks, including things such as content generation, translation, and question answering (QA). However, just like any new technology, LLMs faced some challenges. One of the key issues with LLMs is what’s known as “hallucination.” This happens when the model produces information that is incorrect or made up but still sounds plausible. In this paper, the goal is to outline a framework to help identify and assess these hallucinations through curating new hallucination evaluation methods, datasets and evaluation metrics. The framework is adaptable and can be used with a variety of models and LLMs prod- uct. The main objective is to offer developers and engineers a consistent approach to identifying hallucinations in LLM-based applications before they are released. Bachelor's degree 2024-11-04T01:10:23Z 2024-11-04T01:10:23Z 2024 Final Year Project (FYP) Pan, J. S. H. (2024). Framework to evaluate and test defences against hallucination in large language model. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/180892 https://hdl.handle.net/10356/180892 en application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
Large language model
LLM safety
LLM testing and defences
spellingShingle Computer and Information Science
Large language model
LLM safety
LLM testing and defences
Pan, Johnny Shi Han
Framework to evaluate and test defences against hallucination in large language model
description The recent advancement of AI, particularly the large language models (LLMs) has en- abled unprecedented capabilities in natural language processing (NLP) tasks, including things such as content generation, translation, and question answering (QA). However, just like any new technology, LLMs faced some challenges. One of the key issues with LLMs is what’s known as “hallucination.” This happens when the model produces information that is incorrect or made up but still sounds plausible. In this paper, the goal is to outline a framework to help identify and assess these hallucinations through curating new hallucination evaluation methods, datasets and evaluation metrics. The framework is adaptable and can be used with a variety of models and LLMs prod- uct. The main objective is to offer developers and engineers a consistent approach to identifying hallucinations in LLM-based applications before they are released.
author2 Luu Anh Tuan
author_facet Luu Anh Tuan
Pan, Johnny Shi Han
format Final Year Project
author Pan, Johnny Shi Han
author_sort Pan, Johnny Shi Han
title Framework to evaluate and test defences against hallucination in large language model
title_short Framework to evaluate and test defences against hallucination in large language model
title_full Framework to evaluate and test defences against hallucination in large language model
title_fullStr Framework to evaluate and test defences against hallucination in large language model
title_full_unstemmed Framework to evaluate and test defences against hallucination in large language model
title_sort framework to evaluate and test defences against hallucination in large language model
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/180892
_version_ 1816859006574002176