Robust prompt optimization for large language models against distribution shifts

Large Language Model (LLM) has demonstrated significant ability in various Natural Language Processing tasks. However, their effectiveness is highly dependent on the phrasing of the task prompt, leading to research on automatic prompt optimization using labeled task data. We reveal that these prompt...

Full description

Saved in:
Bibliographic Details
Main Authors: LI, Moxin, WANG, Wenjie, FENG, Fuli, CAO, Yixin, ZHANG, Jizhi, CHUA, Tat-Seng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8393
https://ink.library.smu.edu.sg/context/sis_research/article/9396/viewcontent/2023.emnlp_main.95.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9396
record_format dspace
spelling sg-smu-ink.sis_research-93962024-04-18T02:15:31Z Robust prompt optimization for large language models against distribution shifts LI, Moxin WANG, Wenjie FENG, Fuli CAO, Yixin ZHANG, Jizhi CHUA, Tat-Seng Large Language Model (LLM) has demonstrated significant ability in various Natural Language Processing tasks. However, their effectiveness is highly dependent on the phrasing of the task prompt, leading to research on automatic prompt optimization using labeled task data. We reveal that these prompt optimization techniques are vulnerable to distribution shifts such as subpopulation shifts, which are common for LLMs in real-world scenarios such as customer reviews analysis. In this light, we propose a new problem of robust prompt optimization for LLMs against distribution shifts, which requires the prompt optimized over the labeled source group can simultaneously generalize to an unlabeled target group. To solve this problem, we propose Generalized Prompt Optimization framework , which incorporates the unlabeled data from the target group into prompt optimization. Extensive experimental results demonstrate the effectiveness of the proposed framework with significant performance improvement on the target group and comparable performance on the source group. 2023-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8393 info:doi/10.18653/v1/2023.emnlp-main.95 https://ink.library.smu.edu.sg/context/sis_research/article/9396/viewcontent/2023.emnlp_main.95.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Language model Language processing Natural language Optimizations Optimization framework Artificial Intelligence and Robotics Databases and Information Systems
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Language model
Language processing
Natural language
Optimizations
Optimization framework
Artificial Intelligence and Robotics
Databases and Information Systems
spellingShingle Language model
Language processing
Natural language
Optimizations
Optimization framework
Artificial Intelligence and Robotics
Databases and Information Systems
LI, Moxin
WANG, Wenjie
FENG, Fuli
CAO, Yixin
ZHANG, Jizhi
CHUA, Tat-Seng
Robust prompt optimization for large language models against distribution shifts
description Large Language Model (LLM) has demonstrated significant ability in various Natural Language Processing tasks. However, their effectiveness is highly dependent on the phrasing of the task prompt, leading to research on automatic prompt optimization using labeled task data. We reveal that these prompt optimization techniques are vulnerable to distribution shifts such as subpopulation shifts, which are common for LLMs in real-world scenarios such as customer reviews analysis. In this light, we propose a new problem of robust prompt optimization for LLMs against distribution shifts, which requires the prompt optimized over the labeled source group can simultaneously generalize to an unlabeled target group. To solve this problem, we propose Generalized Prompt Optimization framework , which incorporates the unlabeled data from the target group into prompt optimization. Extensive experimental results demonstrate the effectiveness of the proposed framework with significant performance improvement on the target group and comparable performance on the source group.
format text
author LI, Moxin
WANG, Wenjie
FENG, Fuli
CAO, Yixin
ZHANG, Jizhi
CHUA, Tat-Seng
author_facet LI, Moxin
WANG, Wenjie
FENG, Fuli
CAO, Yixin
ZHANG, Jizhi
CHUA, Tat-Seng
author_sort LI, Moxin
title Robust prompt optimization for large language models against distribution shifts
title_short Robust prompt optimization for large language models against distribution shifts
title_full Robust prompt optimization for large language models against distribution shifts
title_fullStr Robust prompt optimization for large language models against distribution shifts
title_full_unstemmed Robust prompt optimization for large language models against distribution shifts
title_sort robust prompt optimization for large language models against distribution shifts
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/sis_research/8393
https://ink.library.smu.edu.sg/context/sis_research/article/9396/viewcontent/2023.emnlp_main.95.pdf
_version_ 1814047492216455168