Chain of preference optimization: Improving chain-of-thought reasoning in LLMs
The recent development of chain-of-thought (CoT) decoding has enabled large language models (LLMs) to generate explicit logical reasoning paths for complex problem-solving. However, research indicates that these paths are not always deliberate and optimal. The tree-of-thought (ToT) method employs tr...
Saved in:
Main Authors: | ZHANG, Xuan, DU, Chao, PANG, Tianyu, LIU, Qian, GAO, Wei, LIN, Min |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2024
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/9881 https://ink.library.smu.edu.sg/context/sis_research/article/10881/viewcontent/2406.09136v2.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
Similar Items
-
Cue-CoT: Chain-of-thought prompting for responding to in-depth dialogue questions with LLMs
by: WANG, Hongru, et al.
Published: (2023) -
Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models
by: WANG, Lei, et al.
Published: (2023) -
Multimodal misinformation detection by learning from synthetic data with multimodal LLMs
by: ZENG, Fengzhu, et al.
Published: (2024) -
T-SciQ: Teaching multimodal Chain-of-Thought reasoning via large language model signals for science question answering
by: WANG, Lei, et al.
Published: (2024) -
Exploiting Reasoning Chains for Multi-hop Science Question Answering
by: XU, Weiwen, et al.
Published: (2021)