Hallucination detection: Robustly discerning reliable answers in Large Language Models

Large language models (LLMs) have gained widespread adoption in various natural language processing tasks, including question answering and dialogue systems. However, a major drawback of LLMs is the issue of hallucination, where they generate unfaithful or inconsistent content that deviates from the...

Full description

Saved in:
Bibliographic Details
Main Authors: CHEN, Yuyuan, FU, Qiang, YUAN, Yichen, WEN, Zhihao, FAN, Ge, LIU, Dayiheng, ZHANG, Dongmei, LI, Zhixu, XIAO, Yanghua
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8464
https://ink.library.smu.edu.sg/context/sis_research/article/9467/viewcontent/3583780.3614905_pv.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9467
record_format dspace
spelling sg-smu-ink.sis_research-94672024-01-04T09:42:25Z Hallucination detection: Robustly discerning reliable answers in Large Language Models CHEN, Yuyuan FU, Qiang YUAN, Yichen WEN, Zhihao FAN, Ge LIU, Dayiheng ZHANG, Dongmei LI, Zhixu XIAO, Yanghua Large language models (LLMs) have gained widespread adoption in various natural language processing tasks, including question answering and dialogue systems. However, a major drawback of LLMs is the issue of hallucination, where they generate unfaithful or inconsistent content that deviates from the input source, leading to severe consequences. In this paper, we propose a robust discriminator named RelD to effectively detect hallucination in LLMs' generated answers. RelD is trained on the constructed RelQA, a bilingual question-answering dialogue dataset along with answers generated by LLMs and a comprehensive set of metrics. Our experimental results demonstrate that the proposed RelD successfully detects hallucination in the answers generated by diverse LLMs. Moreover, it performs well in distinguishing hallucination in LLMs' generated answers from both in-distribution and out-of-distribution datasets. Additionally, we also conduct a thorough analysis of the types of hallucinations that occur and present valuable insights. This research significantly contributes to the detection of reliable answers generated by LLMs and holds noteworthy implications for mitigating hallucination in the future work. 2023-10-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8464 info:doi/10.1145/3583780.3614905 https://ink.library.smu.edu.sg/context/sis_research/article/9467/viewcontent/3583780.3614905_pv.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Hallucination Detection Large Language Models Reliable Answers Artificial Intelligence and Robotics Numerical Analysis and Scientific Computing Programming Languages and Compilers
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Hallucination Detection
Large Language Models
Reliable Answers
Artificial Intelligence and Robotics
Numerical Analysis and Scientific Computing
Programming Languages and Compilers
spellingShingle Hallucination Detection
Large Language Models
Reliable Answers
Artificial Intelligence and Robotics
Numerical Analysis and Scientific Computing
Programming Languages and Compilers
CHEN, Yuyuan
FU, Qiang
YUAN, Yichen
WEN, Zhihao
FAN, Ge
LIU, Dayiheng
ZHANG, Dongmei
LI, Zhixu
XIAO, Yanghua
Hallucination detection: Robustly discerning reliable answers in Large Language Models
description Large language models (LLMs) have gained widespread adoption in various natural language processing tasks, including question answering and dialogue systems. However, a major drawback of LLMs is the issue of hallucination, where they generate unfaithful or inconsistent content that deviates from the input source, leading to severe consequences. In this paper, we propose a robust discriminator named RelD to effectively detect hallucination in LLMs' generated answers. RelD is trained on the constructed RelQA, a bilingual question-answering dialogue dataset along with answers generated by LLMs and a comprehensive set of metrics. Our experimental results demonstrate that the proposed RelD successfully detects hallucination in the answers generated by diverse LLMs. Moreover, it performs well in distinguishing hallucination in LLMs' generated answers from both in-distribution and out-of-distribution datasets. Additionally, we also conduct a thorough analysis of the types of hallucinations that occur and present valuable insights. This research significantly contributes to the detection of reliable answers generated by LLMs and holds noteworthy implications for mitigating hallucination in the future work.
format text
author CHEN, Yuyuan
FU, Qiang
YUAN, Yichen
WEN, Zhihao
FAN, Ge
LIU, Dayiheng
ZHANG, Dongmei
LI, Zhixu
XIAO, Yanghua
author_facet CHEN, Yuyuan
FU, Qiang
YUAN, Yichen
WEN, Zhihao
FAN, Ge
LIU, Dayiheng
ZHANG, Dongmei
LI, Zhixu
XIAO, Yanghua
author_sort CHEN, Yuyuan
title Hallucination detection: Robustly discerning reliable answers in Large Language Models
title_short Hallucination detection: Robustly discerning reliable answers in Large Language Models
title_full Hallucination detection: Robustly discerning reliable answers in Large Language Models
title_fullStr Hallucination detection: Robustly discerning reliable answers in Large Language Models
title_full_unstemmed Hallucination detection: Robustly discerning reliable answers in Large Language Models
title_sort hallucination detection: robustly discerning reliable answers in large language models
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/sis_research/8464
https://ink.library.smu.edu.sg/context/sis_research/article/9467/viewcontent/3583780.3614905_pv.pdf
_version_ 1787590774440853504