Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models

Large language models (LLMs) have recently been shown to deliver impressive performance in various NLP tasks. To tackle multi-step reasoning tasks, few-shot chain-of-thought (CoT) prompting includes a few manually crafted step-by-step reasoning demonstrations which enable LLMs to explicitly generate...

Full description

Saved in:
Bibliographic Details
Main Authors: WANG, Lei, XU, Wanyu, LAN, Yihuai, HU, Zhiqiang, LAN, Yunshi, LEE, Roy Ka-Wei, LIM, Ee-peng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8054
https://ink.library.smu.edu.sg/context/sis_research/article/9057/viewcontent/2023.acl_long.147.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9057
record_format dspace
spelling sg-smu-ink.sis_research-90572024-04-18T05:36:23Z Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models WANG, Lei XU, Wanyu LAN, Yihuai HU, Zhiqiang LAN, Yunshi LEE, Roy Ka-Wei LIM, Ee-peng Large language models (LLMs) have recently been shown to deliver impressive performance in various NLP tasks. To tackle multi-step reasoning tasks, few-shot chain-of-thought (CoT) prompting includes a few manually crafted step-by-step reasoning demonstrations which enable LLMs to explicitly generate reasoning steps and improve their reasoning task accuracy. To eliminate the manual effort, Zeroshot-CoT concatenates the target problem statement with “Let’s think step by step” as an input prompt to LLMs. Despite the success of Zero-shot-CoT, it still suffers from three pitfalls: calculation errors, missing-step errors, and semantic misunderstanding errors. To address the missing-step errors, we propose Planand-Solve (PS) Prompting. It consists of two components: first, devising a plan to divide the entire task into smaller subtasks, and then carrying out the subtasks according to the plan. To address the calculation errors and improve the quality of generated reasoning steps, we extend PS prompting with more detailed instructions and derive PS+ prompting. We evaluate our proposed prompting strategy on ten datasets across three reasoning problems. The experimental results over GPT-3 show that our proposed zero-shot prompting consistently outperforms Zero-shot-CoT across all datasets by a large margin, is comparable to or exceeds Zero-shot-Program-of-Thought Prompting, and has comparable performance with 8-shot CoT prompting on the math reasoning problem. The code can be found at https://github.com/AGIEdgerunners/Plan-and-Solve-Prompting. 2023-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8054 info:doi/10.18653/v1/2023.acl-long.147 https://ink.library.smu.edu.sg/context/sis_research/article/9057/viewcontent/2023.acl_long.147.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Calculation error Language model Large margins Multisteps Performance Reasoning problems Artificial Intelligence and Robotics Databases and Information Systems Programming Languages and Compilers
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Calculation error
Language model
Large margins
Multisteps
Performance
Reasoning problems
Artificial Intelligence and Robotics
Databases and Information Systems
Programming Languages and Compilers
spellingShingle Calculation error
Language model
Large margins
Multisteps
Performance
Reasoning problems
Artificial Intelligence and Robotics
Databases and Information Systems
Programming Languages and Compilers
WANG, Lei
XU, Wanyu
LAN, Yihuai
HU, Zhiqiang
LAN, Yunshi
LEE, Roy Ka-Wei
LIM, Ee-peng
Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models
description Large language models (LLMs) have recently been shown to deliver impressive performance in various NLP tasks. To tackle multi-step reasoning tasks, few-shot chain-of-thought (CoT) prompting includes a few manually crafted step-by-step reasoning demonstrations which enable LLMs to explicitly generate reasoning steps and improve their reasoning task accuracy. To eliminate the manual effort, Zeroshot-CoT concatenates the target problem statement with “Let’s think step by step” as an input prompt to LLMs. Despite the success of Zero-shot-CoT, it still suffers from three pitfalls: calculation errors, missing-step errors, and semantic misunderstanding errors. To address the missing-step errors, we propose Planand-Solve (PS) Prompting. It consists of two components: first, devising a plan to divide the entire task into smaller subtasks, and then carrying out the subtasks according to the plan. To address the calculation errors and improve the quality of generated reasoning steps, we extend PS prompting with more detailed instructions and derive PS+ prompting. We evaluate our proposed prompting strategy on ten datasets across three reasoning problems. The experimental results over GPT-3 show that our proposed zero-shot prompting consistently outperforms Zero-shot-CoT across all datasets by a large margin, is comparable to or exceeds Zero-shot-Program-of-Thought Prompting, and has comparable performance with 8-shot CoT prompting on the math reasoning problem. The code can be found at https://github.com/AGIEdgerunners/Plan-and-Solve-Prompting.
format text
author WANG, Lei
XU, Wanyu
LAN, Yihuai
HU, Zhiqiang
LAN, Yunshi
LEE, Roy Ka-Wei
LIM, Ee-peng
author_facet WANG, Lei
XU, Wanyu
LAN, Yihuai
HU, Zhiqiang
LAN, Yunshi
LEE, Roy Ka-Wei
LIM, Ee-peng
author_sort WANG, Lei
title Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models
title_short Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models
title_full Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models
title_fullStr Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models
title_full_unstemmed Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models
title_sort plan-and-solve prompting: improving zero-shot chain-of-thought reasoning by large language models
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/sis_research/8054
https://ink.library.smu.edu.sg/context/sis_research/article/9057/viewcontent/2023.acl_long.147.pdf
_version_ 1814047491508666368