Hallucination detection: Robustly discerning reliable answers in Large Language Models
Large language models (LLMs) have gained widespread adoption in various natural language processing tasks, including question answering and dialogue systems. However, a major drawback of LLMs is the issue of hallucination, where they generate unfaithful or inconsistent content that deviates from the...
Saved in:
Main Authors: | , , , , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2023
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8464 https://ink.library.smu.edu.sg/context/sis_research/article/9467/viewcontent/3583780.3614905_pv.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9467 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-94672024-01-04T09:42:25Z Hallucination detection: Robustly discerning reliable answers in Large Language Models CHEN, Yuyuan FU, Qiang YUAN, Yichen WEN, Zhihao FAN, Ge LIU, Dayiheng ZHANG, Dongmei LI, Zhixu XIAO, Yanghua Large language models (LLMs) have gained widespread adoption in various natural language processing tasks, including question answering and dialogue systems. However, a major drawback of LLMs is the issue of hallucination, where they generate unfaithful or inconsistent content that deviates from the input source, leading to severe consequences. In this paper, we propose a robust discriminator named RelD to effectively detect hallucination in LLMs' generated answers. RelD is trained on the constructed RelQA, a bilingual question-answering dialogue dataset along with answers generated by LLMs and a comprehensive set of metrics. Our experimental results demonstrate that the proposed RelD successfully detects hallucination in the answers generated by diverse LLMs. Moreover, it performs well in distinguishing hallucination in LLMs' generated answers from both in-distribution and out-of-distribution datasets. Additionally, we also conduct a thorough analysis of the types of hallucinations that occur and present valuable insights. This research significantly contributes to the detection of reliable answers generated by LLMs and holds noteworthy implications for mitigating hallucination in the future work. 2023-10-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8464 info:doi/10.1145/3583780.3614905 https://ink.library.smu.edu.sg/context/sis_research/article/9467/viewcontent/3583780.3614905_pv.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Hallucination Detection Large Language Models Reliable Answers Artificial Intelligence and Robotics Numerical Analysis and Scientific Computing Programming Languages and Compilers |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Hallucination Detection Large Language Models Reliable Answers Artificial Intelligence and Robotics Numerical Analysis and Scientific Computing Programming Languages and Compilers |
spellingShingle |
Hallucination Detection Large Language Models Reliable Answers Artificial Intelligence and Robotics Numerical Analysis and Scientific Computing Programming Languages and Compilers CHEN, Yuyuan FU, Qiang YUAN, Yichen WEN, Zhihao FAN, Ge LIU, Dayiheng ZHANG, Dongmei LI, Zhixu XIAO, Yanghua Hallucination detection: Robustly discerning reliable answers in Large Language Models |
description |
Large language models (LLMs) have gained widespread adoption in various natural language processing tasks, including question answering and dialogue systems. However, a major drawback of LLMs is the issue of hallucination, where they generate unfaithful or inconsistent content that deviates from the input source, leading to severe consequences. In this paper, we propose a robust discriminator named RelD to effectively detect hallucination in LLMs' generated answers. RelD is trained on the constructed RelQA, a bilingual question-answering dialogue dataset along with answers generated by LLMs and a comprehensive set of metrics. Our experimental results demonstrate that the proposed RelD successfully detects hallucination in the answers generated by diverse LLMs. Moreover, it performs well in distinguishing hallucination in LLMs' generated answers from both in-distribution and out-of-distribution datasets. Additionally, we also conduct a thorough analysis of the types of hallucinations that occur and present valuable insights. This research significantly contributes to the detection of reliable answers generated by LLMs and holds noteworthy implications for mitigating hallucination in the future work. |
format |
text |
author |
CHEN, Yuyuan FU, Qiang YUAN, Yichen WEN, Zhihao FAN, Ge LIU, Dayiheng ZHANG, Dongmei LI, Zhixu XIAO, Yanghua |
author_facet |
CHEN, Yuyuan FU, Qiang YUAN, Yichen WEN, Zhihao FAN, Ge LIU, Dayiheng ZHANG, Dongmei LI, Zhixu XIAO, Yanghua |
author_sort |
CHEN, Yuyuan |
title |
Hallucination detection: Robustly discerning reliable answers in Large Language Models |
title_short |
Hallucination detection: Robustly discerning reliable answers in Large Language Models |
title_full |
Hallucination detection: Robustly discerning reliable answers in Large Language Models |
title_fullStr |
Hallucination detection: Robustly discerning reliable answers in Large Language Models |
title_full_unstemmed |
Hallucination detection: Robustly discerning reliable answers in Large Language Models |
title_sort |
hallucination detection: robustly discerning reliable answers in large language models |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2023 |
url |
https://ink.library.smu.edu.sg/sis_research/8464 https://ink.library.smu.edu.sg/context/sis_research/article/9467/viewcontent/3583780.3614905_pv.pdf |
_version_ |
1787590774440853504 |