Towards robust personalized dialogue generation via order-insensitive representation regularization

Generating persona consistent dialogue response is important for developing an intelligent conversational agent. Recent works typically fine-tune large-scale pre-trained models on this task by concatenating persona texts and dialogue history as a single input sequence to generate the target response...

Full description

Saved in:
Bibliographic Details
Main Authors: CHEN, Liang, WANG, Hongru, DENG, Yang, KWAN, Wai-Chung, WANG, Zezhong, WONG, Kam-Fai
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/9125
https://ink.library.smu.edu.sg/context/sis_research/article/10128/viewcontent/2023.findings_acl.462.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-10128
record_format dspace
spelling sg-smu-ink.sis_research-101282024-08-01T14:27:14Z Towards robust personalized dialogue generation via order-insensitive representation regularization CHEN, Liang WANG, Hongru DENG, Yang KWAN, Wai-Chung WANG, Zezhong WONG, Kam-Fai Generating persona consistent dialogue response is important for developing an intelligent conversational agent. Recent works typically fine-tune large-scale pre-trained models on this task by concatenating persona texts and dialogue history as a single input sequence to generate the target response. While simple and effective, our analysis shows that this popular practice is seriously affected by order sensitivity where different input orders of persona sentences significantly impact the quality and consistency of generated response, resulting in severe performance fluctuations (i.e., 29.4% on GPT2 and 83.2% on BART). To mitigate the order sensitivity problem, we propose a model-agnostic framework, ORder Insensitive Generation (ORIG), which enables dialogue models to learn robust representation under different persona orders and improve the consistency of response generation. Experiments on the Persona-Chat dataset justify the effectiveness and superiority of our method with two dominant pre-trained models (GPT2 and BART). 2023-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9125 info:doi/10.18653/v1/2023.findings-acl.462 https://ink.library.smu.edu.sg/context/sis_research/article/10128/viewcontent/2023.findings_acl.462.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Databases and Information Systems
spellingShingle Databases and Information Systems
CHEN, Liang
WANG, Hongru
DENG, Yang
KWAN, Wai-Chung
WANG, Zezhong
WONG, Kam-Fai
Towards robust personalized dialogue generation via order-insensitive representation regularization
description Generating persona consistent dialogue response is important for developing an intelligent conversational agent. Recent works typically fine-tune large-scale pre-trained models on this task by concatenating persona texts and dialogue history as a single input sequence to generate the target response. While simple and effective, our analysis shows that this popular practice is seriously affected by order sensitivity where different input orders of persona sentences significantly impact the quality and consistency of generated response, resulting in severe performance fluctuations (i.e., 29.4% on GPT2 and 83.2% on BART). To mitigate the order sensitivity problem, we propose a model-agnostic framework, ORder Insensitive Generation (ORIG), which enables dialogue models to learn robust representation under different persona orders and improve the consistency of response generation. Experiments on the Persona-Chat dataset justify the effectiveness and superiority of our method with two dominant pre-trained models (GPT2 and BART).
format text
author CHEN, Liang
WANG, Hongru
DENG, Yang
KWAN, Wai-Chung
WANG, Zezhong
WONG, Kam-Fai
author_facet CHEN, Liang
WANG, Hongru
DENG, Yang
KWAN, Wai-Chung
WANG, Zezhong
WONG, Kam-Fai
author_sort CHEN, Liang
title Towards robust personalized dialogue generation via order-insensitive representation regularization
title_short Towards robust personalized dialogue generation via order-insensitive representation regularization
title_full Towards robust personalized dialogue generation via order-insensitive representation regularization
title_fullStr Towards robust personalized dialogue generation via order-insensitive representation regularization
title_full_unstemmed Towards robust personalized dialogue generation via order-insensitive representation regularization
title_sort towards robust personalized dialogue generation via order-insensitive representation regularization
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/sis_research/9125
https://ink.library.smu.edu.sg/context/sis_research/article/10128/viewcontent/2023.findings_acl.462.pdf
_version_ 1814047749175246848