MITIGATING HALLUCINATION IN AUTOMATED INTERVIEWS USING A CHAIN-OF-VERIFICATION APPROACH ON LLMS
The use of Large Language Models (LLMs) in recruitment processes presents significant challenges, primarily due to the inherent tendency of LLMs to produce hallucinations. This research develops an anti-hallucination component based on the chain-of-verification method, utilizing GPT-3.5 and GPT-4...
Saved in:
Main Author: | |
---|---|
Format: | Final Project |
Language: | Indonesia |
Online Access: | https://digilib.itb.ac.id/gdl/view/85547 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Institut Teknologi Bandung |
Language: | Indonesia |
id |
id-itb.:85547 |
---|---|
spelling |
id-itb.:855472024-08-21T18:07:18ZMITIGATING HALLUCINATION IN AUTOMATED INTERVIEWS USING A CHAIN-OF-VERIFICATION APPROACH ON LLMS Bintang Nurmansyah, Ilham Indonesia Final Project anti-hallucination, chain-of-verification, LLM verifier, GPT, few-shot, LLM interviewer INSTITUT TEKNOLOGI BANDUNG https://digilib.itb.ac.id/gdl/view/85547 The use of Large Language Models (LLMs) in recruitment processes presents significant challenges, primarily due to the inherent tendency of LLMs to produce hallucinations. This research develops an anti-hallucination component based on the chain-of-verification method, utilizing GPT-3.5 and GPT-4 models to detect and prevent LLM interviewers from delivering hallucinated responses to candidates. GPT models were selected due to their lower hallucination rates compared to other models at the time of this study (GPT-3.5 at 3.5% and GPT-4 at 3%). The study employs two GPT models to compare the effectiveness of hallucination prevention. The system implementation uses the chain-of-verification technique, applying few-shot learning to the LLM verifier model to identify and differentiate potential hallucinations. This method is integrated as a module within the LLM interviewer, which is invoked each time the LLM interviewer generates a question or response to a candidate's prompt. The results demonstrate that interviews without the chain-of-verification method exhibit an average hallucination rate of 100% when given certain prompts, whereas, with the chain- of-verification, the average hallucination rate is reduced to 4% under the same conditions. GPT-4 emerges as the most effective model, with an average hallucination rate of 12%, compared to 72% for GPT-3.5. text |
institution |
Institut Teknologi Bandung |
building |
Institut Teknologi Bandung Library |
continent |
Asia |
country |
Indonesia Indonesia |
content_provider |
Institut Teknologi Bandung |
collection |
Digital ITB |
language |
Indonesia |
description |
The use of Large Language Models (LLMs) in recruitment processes presents
significant challenges, primarily due to the inherent tendency of LLMs to produce
hallucinations. This research develops an anti-hallucination component based on
the chain-of-verification method, utilizing GPT-3.5 and GPT-4 models to detect
and prevent LLM interviewers from delivering hallucinated responses to
candidates. GPT models were selected due to their lower hallucination rates
compared to other models at the time of this study (GPT-3.5 at 3.5% and GPT-4 at
3%). The study employs two GPT models to compare the effectiveness of
hallucination prevention. The system implementation uses the chain-of-verification
technique, applying few-shot learning to the LLM verifier model to identify and
differentiate potential hallucinations. This method is integrated as a module within
the LLM interviewer, which is invoked each time the LLM interviewer generates a
question or response to a candidate's prompt. The results demonstrate that
interviews without the chain-of-verification method exhibit an average
hallucination rate of 100% when given certain prompts, whereas, with the chain-
of-verification, the average hallucination rate is reduced to 4% under the same
conditions. GPT-4 emerges as the most effective model, with an average
hallucination rate of 12%, compared to 72% for GPT-3.5. |
format |
Final Project |
author |
Bintang Nurmansyah, Ilham |
spellingShingle |
Bintang Nurmansyah, Ilham MITIGATING HALLUCINATION IN AUTOMATED INTERVIEWS USING A CHAIN-OF-VERIFICATION APPROACH ON LLMS |
author_facet |
Bintang Nurmansyah, Ilham |
author_sort |
Bintang Nurmansyah, Ilham |
title |
MITIGATING HALLUCINATION IN AUTOMATED INTERVIEWS USING A CHAIN-OF-VERIFICATION APPROACH ON LLMS |
title_short |
MITIGATING HALLUCINATION IN AUTOMATED INTERVIEWS USING A CHAIN-OF-VERIFICATION APPROACH ON LLMS |
title_full |
MITIGATING HALLUCINATION IN AUTOMATED INTERVIEWS USING A CHAIN-OF-VERIFICATION APPROACH ON LLMS |
title_fullStr |
MITIGATING HALLUCINATION IN AUTOMATED INTERVIEWS USING A CHAIN-OF-VERIFICATION APPROACH ON LLMS |
title_full_unstemmed |
MITIGATING HALLUCINATION IN AUTOMATED INTERVIEWS USING A CHAIN-OF-VERIFICATION APPROACH ON LLMS |
title_sort |
mitigating hallucination in automated interviews using a chain-of-verification approach on llms |
url |
https://digilib.itb.ac.id/gdl/view/85547 |
_version_ |
1822999207590494208 |