LLMs-as-instructors : Learning from errors toward automating model improvement

This paper introduces the innovative "LLMs-as-Instructors'' framework, which leverages the advanced Large Language Models (LLMs) to autonomously enhance the training of smaller target models. Inspired by the theory of "Learning from Errors'', this framework employs an i...

Full description

Saved in:
Bibliographic Details
Main Authors: YING, Jiahao, LIN, Mingbao, CAO, Yixin, TANG, Wei, WANG, Bo, SUN, Qianru, HUANG, Xuanjing, YAN, Shuicheng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2024
Subjects:
LLM
Online Access:https://ink.library.smu.edu.sg/sis_research/9440
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-10440
record_format dspace
spelling sg-smu-ink.sis_research-104402024-10-24T09:48:03Z LLMs-as-instructors : Learning from errors toward automating model improvement YING, Jiahao LIN, Mingbao CAO, Yixin TANG, Wei WANG, Bo SUN, Qianru HUANG, Xuanjing YAN, Shuicheng This paper introduces the innovative "LLMs-as-Instructors'' framework, which leverages the advanced Large Language Models (LLMs) to autonomously enhance the training of smaller target models. Inspired by the theory of "Learning from Errors'', this framework employs an instructor LLM to meticulously analyze the specific errors within a target model, facilitating targeted and efficient training cycles. Within this framework, we implement two strategies: "Learning from Error," which focuses solely on incorrect responses to tailor training data, and "Learning from Error by Contrast,'' which uses contrastive learning to analyze both correct and incorrect responses for a deeper understanding of errors.Our empirical studies, conducted with several open-source models, demonstrate significant improvements across multiple benchmarks, including mathematical reasoning, coding abilities, and factual knowledge. Notably, the refined Llama-3-8b-Instruction has outperformed ChatGPT, illustrating the effectiveness of our approach. By leveraging the strengths of both strategies, we have attained a more balanced performance improvement on both in-domain and out-of-domain benchmarks. 2024-11-12T08:00:00Z text https://ink.library.smu.edu.sg/sis_research/9440 info:doi/10.48550/arXiv.2407.00497 Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Large Language Models LLM Learning from errors Model training Artificial Intelligence and Robotics
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Large Language Models
LLM
Learning from errors
Model training
Artificial Intelligence and Robotics
spellingShingle Large Language Models
LLM
Learning from errors
Model training
Artificial Intelligence and Robotics
YING, Jiahao
LIN, Mingbao
CAO, Yixin
TANG, Wei
WANG, Bo
SUN, Qianru
HUANG, Xuanjing
YAN, Shuicheng
LLMs-as-instructors : Learning from errors toward automating model improvement
description This paper introduces the innovative "LLMs-as-Instructors'' framework, which leverages the advanced Large Language Models (LLMs) to autonomously enhance the training of smaller target models. Inspired by the theory of "Learning from Errors'', this framework employs an instructor LLM to meticulously analyze the specific errors within a target model, facilitating targeted and efficient training cycles. Within this framework, we implement two strategies: "Learning from Error," which focuses solely on incorrect responses to tailor training data, and "Learning from Error by Contrast,'' which uses contrastive learning to analyze both correct and incorrect responses for a deeper understanding of errors.Our empirical studies, conducted with several open-source models, demonstrate significant improvements across multiple benchmarks, including mathematical reasoning, coding abilities, and factual knowledge. Notably, the refined Llama-3-8b-Instruction has outperformed ChatGPT, illustrating the effectiveness of our approach. By leveraging the strengths of both strategies, we have attained a more balanced performance improvement on both in-domain and out-of-domain benchmarks.
format text
author YING, Jiahao
LIN, Mingbao
CAO, Yixin
TANG, Wei
WANG, Bo
SUN, Qianru
HUANG, Xuanjing
YAN, Shuicheng
author_facet YING, Jiahao
LIN, Mingbao
CAO, Yixin
TANG, Wei
WANG, Bo
SUN, Qianru
HUANG, Xuanjing
YAN, Shuicheng
author_sort YING, Jiahao
title LLMs-as-instructors : Learning from errors toward automating model improvement
title_short LLMs-as-instructors : Learning from errors toward automating model improvement
title_full LLMs-as-instructors : Learning from errors toward automating model improvement
title_fullStr LLMs-as-instructors : Learning from errors toward automating model improvement
title_full_unstemmed LLMs-as-instructors : Learning from errors toward automating model improvement
title_sort llms-as-instructors : learning from errors toward automating model improvement
publisher Institutional Knowledge at Singapore Management University
publishDate 2024
url https://ink.library.smu.edu.sg/sis_research/9440
_version_ 1814777852531310592