LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models

The success of large language models (LLMs), like GPT-4 and ChatGPT, has led to the development of numerous cost-effective and accessible alternatives that are created by finetuning open-access LLMs with task-specific data (e.g., ChatDoctor) or instruction data (e.g., Alpaca). Among the various fine...

Full description

Saved in:
Bibliographic Details
Main Authors: HU, Zhiqiang, WANG, Lei, LAN, Yihuai, XU, Wanyu, LIM, Ee-peng, BING, Lidong, XU, Xing, PORIA, Soujanya, LEE, Roy Ka-Wei
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8324
https://ink.library.smu.edu.sg/context/sis_research/article/9327/viewcontent/2304.01933.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9327
record_format dspace
spelling sg-smu-ink.sis_research-93272023-12-05T03:04:18Z LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models HU, Zhiqiang WANG, Lei LAN, Yihuai XU, Wanyu LIM, Ee-peng BING, Lidong XU, Xing PORIA, Soujanya LEE, Roy Ka-Wei The success of large language models (LLMs), like GPT-4 and ChatGPT, has led to the development of numerous cost-effective and accessible alternatives that are created by finetuning open-access LLMs with task-specific data (e.g., ChatDoctor) or instruction data (e.g., Alpaca). Among the various fine-tuning methods, adapter-based parameter-efficient fine-tuning (PEFT) is undoubtedly one of the most attractive topics, as it only requires fine-tuning a few external parameters instead of the entire LLMs while achieving comparable or even better performance. To enable further research on PEFT methods of LLMs, this paper presents LLMAdapters, an easy-to-use framework that integrates various adapters into LLMs and can execute these adapter-based PEFT methods of LLMs for different tasks. The framework includes state-of-the-art open-access LLMs such as LLaMA, BLOOM, and GPT-J, as well as widely used adapters such as Series adapters, Parallel adapter, Prompt-based learning and Reparametrization-based methods. Moreover, we conduct extensive empirical studies on the impact of adapter types, placement locations, and hyper-parameters to the best design for each adapter-based methods. We evaluate the effectiveness of the adapters on fourteen datasets from two different reasoning tasks, Arithmetic Reasoning and Commonsense Reasoning. The results demonstrate that using adapter-based PEFT in smaller-scale LLMs (7B) with few extra trainable parameters yields comparable, and in some cases superior, performance to powerful LLMs (175B) in zero-shot inference on both reasoning tasks. 2023-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8324 https://ink.library.smu.edu.sg/context/sis_research/article/9327/viewcontent/2304.01933.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems Programming Languages and Compilers
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Databases and Information Systems
Programming Languages and Compilers
spellingShingle Databases and Information Systems
Programming Languages and Compilers
HU, Zhiqiang
WANG, Lei
LAN, Yihuai
XU, Wanyu
LIM, Ee-peng
BING, Lidong
XU, Xing
PORIA, Soujanya
LEE, Roy Ka-Wei
LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models
description The success of large language models (LLMs), like GPT-4 and ChatGPT, has led to the development of numerous cost-effective and accessible alternatives that are created by finetuning open-access LLMs with task-specific data (e.g., ChatDoctor) or instruction data (e.g., Alpaca). Among the various fine-tuning methods, adapter-based parameter-efficient fine-tuning (PEFT) is undoubtedly one of the most attractive topics, as it only requires fine-tuning a few external parameters instead of the entire LLMs while achieving comparable or even better performance. To enable further research on PEFT methods of LLMs, this paper presents LLMAdapters, an easy-to-use framework that integrates various adapters into LLMs and can execute these adapter-based PEFT methods of LLMs for different tasks. The framework includes state-of-the-art open-access LLMs such as LLaMA, BLOOM, and GPT-J, as well as widely used adapters such as Series adapters, Parallel adapter, Prompt-based learning and Reparametrization-based methods. Moreover, we conduct extensive empirical studies on the impact of adapter types, placement locations, and hyper-parameters to the best design for each adapter-based methods. We evaluate the effectiveness of the adapters on fourteen datasets from two different reasoning tasks, Arithmetic Reasoning and Commonsense Reasoning. The results demonstrate that using adapter-based PEFT in smaller-scale LLMs (7B) with few extra trainable parameters yields comparable, and in some cases superior, performance to powerful LLMs (175B) in zero-shot inference on both reasoning tasks.
format text
author HU, Zhiqiang
WANG, Lei
LAN, Yihuai
XU, Wanyu
LIM, Ee-peng
BING, Lidong
XU, Xing
PORIA, Soujanya
LEE, Roy Ka-Wei
author_facet HU, Zhiqiang
WANG, Lei
LAN, Yihuai
XU, Wanyu
LIM, Ee-peng
BING, Lidong
XU, Xing
PORIA, Soujanya
LEE, Roy Ka-Wei
author_sort HU, Zhiqiang
title LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models
title_short LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models
title_full LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models
title_fullStr LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models
title_full_unstemmed LLM-adapters: An adapter family for parameter-efficient fine-tuning of large language models
title_sort llm-adapters: an adapter family for parameter-efficient fine-tuning of large language models
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/sis_research/8324
https://ink.library.smu.edu.sg/context/sis_research/article/9327/viewcontent/2304.01933.pdf
_version_ 1784855634568544256