Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery

In Conversational Intent Discovery (CID), Small Language Models (SLMs) struggle with overfitting to familiar intents and fail to label newly discovered ones. This issue stems from their limited grasp of semantic nuances and their intrinsically discriminative framework. Therefore, we propose Synergiz...

Full description

Saved in:
Bibliographic Details
Main Authors: LIANG, Jinggui, LIAO, Lizi, FEI, Hao, JIANG, Jing
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2024
Subjects:
CID
Online Access:https://ink.library.smu.edu.sg/sis_research/9698
https://ink.library.smu.edu.sg/context/sis_research/article/10698/viewcontent/2024.findings_acl.840.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-10698
record_format dspace
spelling sg-smu-ink.sis_research-106982024-11-28T09:03:44Z Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery LIANG, Jinggui LIAO, Lizi FEI, Hao JIANG, Jing In Conversational Intent Discovery (CID), Small Language Models (SLMs) struggle with overfitting to familiar intents and fail to label newly discovered ones. This issue stems from their limited grasp of semantic nuances and their intrinsically discriminative framework. Therefore, we propose Synergizing Large Language Models (LLMs) with pre-trained SLMs for CID (SynCID). It harnesses the profound semantic comprehension of LLMs alongside the operational agility of SLMs. By utilizing LLMs to refine both utterances and existing intent labels, SynCID significantly enhances the semantic depth, subsequently realigning these enriched descriptors within the SLMs’ feature space to correct cluster distortion and promote robust learning of representations. A key advantage is its capacity for the early identification of new intents, a critical aspect for deploying conversational agents successfully. Additionally, SynCID leverages the in-context learning strengths of LLMs to generate labels for new intents. Thorough evaluations across a wide array of datasets have demonstrated its superior performance over traditional CID methods. 2024-08-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9698 info:doi/10.18653/v1/2024.findings-acl.840 https://ink.library.smu.edu.sg/context/sis_research/article/10698/viewcontent/2024.findings_acl.840.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Conversational Intent Discovery CID Large Language Models LLMs Small Language Models SLMs Artificial Intelligence and Robotics Computer Sciences
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Conversational Intent Discovery
CID
Large Language Models
LLMs
Small Language Models
SLMs
Artificial Intelligence and Robotics
Computer Sciences
spellingShingle Conversational Intent Discovery
CID
Large Language Models
LLMs
Small Language Models
SLMs
Artificial Intelligence and Robotics
Computer Sciences
LIANG, Jinggui
LIAO, Lizi
FEI, Hao
JIANG, Jing
Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery
description In Conversational Intent Discovery (CID), Small Language Models (SLMs) struggle with overfitting to familiar intents and fail to label newly discovered ones. This issue stems from their limited grasp of semantic nuances and their intrinsically discriminative framework. Therefore, we propose Synergizing Large Language Models (LLMs) with pre-trained SLMs for CID (SynCID). It harnesses the profound semantic comprehension of LLMs alongside the operational agility of SLMs. By utilizing LLMs to refine both utterances and existing intent labels, SynCID significantly enhances the semantic depth, subsequently realigning these enriched descriptors within the SLMs’ feature space to correct cluster distortion and promote robust learning of representations. A key advantage is its capacity for the early identification of new intents, a critical aspect for deploying conversational agents successfully. Additionally, SynCID leverages the in-context learning strengths of LLMs to generate labels for new intents. Thorough evaluations across a wide array of datasets have demonstrated its superior performance over traditional CID methods.
format text
author LIANG, Jinggui
LIAO, Lizi
FEI, Hao
JIANG, Jing
author_facet LIANG, Jinggui
LIAO, Lizi
FEI, Hao
JIANG, Jing
author_sort LIANG, Jinggui
title Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery
title_short Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery
title_full Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery
title_fullStr Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery
title_full_unstemmed Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery
title_sort synergizing large language models and pre-trained smaller models for conversational intent discovery
publisher Institutional Knowledge at Singapore Management University
publishDate 2024
url https://ink.library.smu.edu.sg/sis_research/9698
https://ink.library.smu.edu.sg/context/sis_research/article/10698/viewcontent/2024.findings_acl.840.pdf
_version_ 1819113106201116672