Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery
In Conversational Intent Discovery (CID), Small Language Models (SLMs) struggle with overfitting to familiar intents and fail to label newly discovered ones. This issue stems from their limited grasp of semantic nuances and their intrinsically discriminative framework. Therefore, we propose Synergiz...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2024
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/9698 https://ink.library.smu.edu.sg/context/sis_research/article/10698/viewcontent/2024.findings_acl.840.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-10698 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-106982024-11-28T09:03:44Z Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery LIANG, Jinggui LIAO, Lizi FEI, Hao JIANG, Jing In Conversational Intent Discovery (CID), Small Language Models (SLMs) struggle with overfitting to familiar intents and fail to label newly discovered ones. This issue stems from their limited grasp of semantic nuances and their intrinsically discriminative framework. Therefore, we propose Synergizing Large Language Models (LLMs) with pre-trained SLMs for CID (SynCID). It harnesses the profound semantic comprehension of LLMs alongside the operational agility of SLMs. By utilizing LLMs to refine both utterances and existing intent labels, SynCID significantly enhances the semantic depth, subsequently realigning these enriched descriptors within the SLMs’ feature space to correct cluster distortion and promote robust learning of representations. A key advantage is its capacity for the early identification of new intents, a critical aspect for deploying conversational agents successfully. Additionally, SynCID leverages the in-context learning strengths of LLMs to generate labels for new intents. Thorough evaluations across a wide array of datasets have demonstrated its superior performance over traditional CID methods. 2024-08-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9698 info:doi/10.18653/v1/2024.findings-acl.840 https://ink.library.smu.edu.sg/context/sis_research/article/10698/viewcontent/2024.findings_acl.840.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Conversational Intent Discovery CID Large Language Models LLMs Small Language Models SLMs Artificial Intelligence and Robotics Computer Sciences |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Conversational Intent Discovery CID Large Language Models LLMs Small Language Models SLMs Artificial Intelligence and Robotics Computer Sciences |
spellingShingle |
Conversational Intent Discovery CID Large Language Models LLMs Small Language Models SLMs Artificial Intelligence and Robotics Computer Sciences LIANG, Jinggui LIAO, Lizi FEI, Hao JIANG, Jing Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery |
description |
In Conversational Intent Discovery (CID), Small Language Models (SLMs) struggle with overfitting to familiar intents and fail to label newly discovered ones. This issue stems from their limited grasp of semantic nuances and their intrinsically discriminative framework. Therefore, we propose Synergizing Large Language Models (LLMs) with pre-trained SLMs for CID (SynCID). It harnesses the profound semantic comprehension of LLMs alongside the operational agility of SLMs. By utilizing LLMs to refine both utterances and existing intent labels, SynCID significantly enhances the semantic depth, subsequently realigning these enriched descriptors within the SLMs’ feature space to correct cluster distortion and promote robust learning of representations. A key advantage is its capacity for the early identification of new intents, a critical aspect for deploying conversational agents successfully. Additionally, SynCID leverages the in-context learning strengths of LLMs to generate labels for new intents. Thorough evaluations across a wide array of datasets have demonstrated its superior performance over traditional CID methods. |
format |
text |
author |
LIANG, Jinggui LIAO, Lizi FEI, Hao JIANG, Jing |
author_facet |
LIANG, Jinggui LIAO, Lizi FEI, Hao JIANG, Jing |
author_sort |
LIANG, Jinggui |
title |
Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery |
title_short |
Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery |
title_full |
Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery |
title_fullStr |
Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery |
title_full_unstemmed |
Synergizing Large Language Models and pre-trained smaller models for conversational intent discovery |
title_sort |
synergizing large language models and pre-trained smaller models for conversational intent discovery |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2024 |
url |
https://ink.library.smu.edu.sg/sis_research/9698 https://ink.library.smu.edu.sg/context/sis_research/article/10698/viewcontent/2024.findings_acl.840.pdf |
_version_ |
1819113106201116672 |