LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models
The success of large language models (LLMs), like GPT-4 and ChatGPT, has led to the development of numerous cost-effective and accessible alternatives that are created by finetuning open-access LLMs with task-specific data (e.g., ChatDoctor) or instruction data (e.g., Alpaca). Among the various fine...
Saved in:
Main Authors: | , , , , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2023
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8324 https://ink.library.smu.edu.sg/context/sis_research/article/9327/viewcontent/2304.01933.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9327 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-93272023-12-05T03:04:18Z LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models HU, Zhiqiang WANG, Lei LAN, Yihuai XU, Wanyu LIM, Ee-peng BING, Lidong XU, Xing PORIA, Soujanya LEE, Roy Ka-Wei The success of large language models (LLMs), like GPT-4 and ChatGPT, has led to the development of numerous cost-effective and accessible alternatives that are created by finetuning open-access LLMs with task-specific data (e.g., ChatDoctor) or instruction data (e.g., Alpaca). Among the various fine-tuning methods, adapter-based parameter-efficient fine-tuning (PEFT) is undoubtedly one of the most attractive topics, as it only requires fine-tuning a few external parameters instead of the entire LLMs while achieving comparable or even better performance. To enable further research on PEFT methods of LLMs, this paper presents LLMAdapters, an easy-to-use framework that integrates various adapters into LLMs and can execute these adapter-based PEFT methods of LLMs for different tasks. The framework includes state-of-the-art open-access LLMs such as LLaMA, BLOOM, and GPT-J, as well as widely used adapters such as Series adapters, Parallel adapter, Prompt-based learning and Reparametrization-based methods. Moreover, we conduct extensive empirical studies on the impact of adapter types, placement locations, and hyper-parameters to the best design for each adapter-based methods. We evaluate the effectiveness of the adapters on fourteen datasets from two different reasoning tasks, Arithmetic Reasoning and Commonsense Reasoning. The results demonstrate that using adapter-based PEFT in smaller-scale LLMs (7B) with few extra trainable parameters yields comparable, and in some cases superior, performance to powerful LLMs (175B) in zero-shot inference on both reasoning tasks. 2023-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8324 https://ink.library.smu.edu.sg/context/sis_research/article/9327/viewcontent/2304.01933.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems Programming Languages and Compilers |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Databases and Information Systems Programming Languages and Compilers |
spellingShingle |
Databases and Information Systems Programming Languages and Compilers HU, Zhiqiang WANG, Lei LAN, Yihuai XU, Wanyu LIM, Ee-peng BING, Lidong XU, Xing PORIA, Soujanya LEE, Roy Ka-Wei LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models |
description |
The success of large language models (LLMs), like GPT-4 and ChatGPT, has led to the development of numerous cost-effective and accessible alternatives that are created by finetuning open-access LLMs with task-specific data (e.g., ChatDoctor) or instruction data (e.g., Alpaca). Among the various fine-tuning methods, adapter-based parameter-efficient fine-tuning (PEFT) is undoubtedly one of the most attractive topics, as it only requires fine-tuning a few external parameters instead of the entire LLMs while achieving comparable or even better performance. To enable further research on PEFT methods of LLMs, this paper presents LLMAdapters, an easy-to-use framework that integrates various adapters into LLMs and can execute these adapter-based PEFT methods of LLMs for different tasks. The framework includes state-of-the-art open-access LLMs such as LLaMA, BLOOM, and GPT-J, as well as widely used adapters such as Series adapters, Parallel adapter, Prompt-based learning and Reparametrization-based methods. Moreover, we conduct extensive empirical studies on the impact of adapter types, placement locations, and hyper-parameters to the best design for each adapter-based methods. We evaluate the effectiveness of the adapters on fourteen datasets from two different reasoning tasks, Arithmetic Reasoning and Commonsense Reasoning. The results demonstrate that using adapter-based PEFT in smaller-scale LLMs (7B) with few extra trainable parameters yields comparable, and in some cases superior, performance to powerful LLMs (175B) in zero-shot inference on both reasoning tasks. |
format |
text |
author |
HU, Zhiqiang WANG, Lei LAN, Yihuai XU, Wanyu LIM, Ee-peng BING, Lidong XU, Xing PORIA, Soujanya LEE, Roy Ka-Wei |
author_facet |
HU, Zhiqiang WANG, Lei LAN, Yihuai XU, Wanyu LIM, Ee-peng BING, Lidong XU, Xing PORIA, Soujanya LEE, Roy Ka-Wei |
author_sort |
HU, Zhiqiang |
title |
LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models |
title_short |
LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models |
title_full |
LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models |
title_fullStr |
LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models |
title_full_unstemmed |
LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models |
title_sort |
llm-adapters: an adapter family for parameter-efficient fine-tuning of large language models |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2023 |
url |
https://ink.library.smu.edu.sg/sis_research/8324 https://ink.library.smu.edu.sg/context/sis_research/article/9327/viewcontent/2304.01933.pdf |
_version_ |
1784855634568544256 |