FUSED DIALOGUE SYSTEM ON OPEN-SOURCE LARGE LANGUAGE MODEL USING END-TO-END APPROACH
In recent years, many researchers have focused on improving dialogue systems in both task-oriented dialogue (ToD) and chitchat domains. However, to create a dialogue system that truly mimics human conversation, we need to combine these objectives, as natural human interactions usually involve bot...
Saved in:
Main Author: | |
---|---|
Format: | Theses |
Language: | Indonesia |
Online Access: | https://digilib.itb.ac.id/gdl/view/84274 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Institut Teknologi Bandung |
Language: | Indonesia |
Summary: | In recent years, many researchers have focused on improving dialogue systems in
both task-oriented dialogue (ToD) and chitchat domains. However, to create a
dialogue system that truly mimics human conversation, we need to combine these
objectives, as natural human interactions usually involve both engaging and
informative elements. So far, there has been no significant work addressing this
challenge using large language models, particularly open-source ones.
This thesis offers a comprehensive exploration of the fused dialogue system task
using an open-source large language model, along with an analysis of the model's
performance and error patterns. Additionally, this thesis proposes a robust
framework to tackle this task using just a belief span 'bpsn'. Our approach opens
new possibilitie and provides several insights for future research in this domain.
The experiments in this thesis include experiments with the proposed mdc-fuse
architecture using Mistral-7B instruct and GODEL-base, the use of naive
approaches in several subtasks such as dialogue state tracking, response
generation, and end-to-end dialogue on two types of open-source LLMs, namely
Mistral-7B instruct and Llama-3, to observe their comparison, and analyzing
compatibility between InstructTODS and smaller open-source LLMs such as
Mistral-7B instruct, followed by several additional analysis in fused dialogue
system.
The experimental results indicate that the choice of language models and training
methods in the mdc-fuse architecture significantly impacts its performance.
Furthermore, the type of open-source LLMs used also plays a critical role.
Specifically, architectures utilizing LLMs like Llama-3 tend to achieve higher score
metrics compared to those using Mistral-7B instruct. Among the cases studied,
mdc-fuse with GODEL-base demonstrated the best performance when compared to
Mistral-7B instruct. Additionally, the experiments reveal that increasing the
number of shots and limiting the dialogue context can further enhance the
architecture's performance. |
---|