INCORPORATING PARAMETER-EFFICIENT FINE-TUNING INTO INDOLEM EVALUATION TASK

The fine-tuning method is employed as a training approach for evaluating various NLU tasks. IndoLEM, which is a pioneer in the evaluation of Indonesian-language NLU, uses fine-tuning as its training method. Fine-tuning involves training the mo- del by modifying all of its parameters, which can...

Full description

Saved in:
Bibliographic Details
Main Author: Prasetya Wicaksana, Adiyansa
Format: Final Project
Language:Indonesia
Online Access:https://digilib.itb.ac.id/gdl/view/85035
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Institut Teknologi Bandung
Language: Indonesia
Description
Summary:The fine-tuning method is employed as a training approach for evaluating various NLU tasks. IndoLEM, which is a pioneer in the evaluation of Indonesian-language NLU, uses fine-tuning as its training method. Fine-tuning involves training the mo- del by modifying all of its parameters, which can be challenging in terms of memory and training time. There exists a method called PEFT that can train models with per- formance comparable to fine-tuning. In this thesis, various PEFT methods, namely LoRA, Prefix-Tuning, Adapter, dan UniPELT, are utilized in the evaluation tasks of IndoLEM. The aim of this thesis is to leverage PEFT methods in IndoLEM, inclu- ding the incorporating of PEFT methods, performance comparisons for each PEFT method, and analysis of parameter usage and training time. This thesis successfu- lly leverages PEFT methods on IndoLEM. Through refactoring of IndoLEM, PEFT methods were successfully incorporated. Subsequently, experiments were condu- cted by training models using both fine-tuning and PEFT methods. Testing was carried out on three evaluation tasks, namely named entity recognition (NER), sen- timent analysis, dan summarization. The experimental results indicate that PEFT only uses approximately 0.2% to 15% of the model’s training parameters, with fas- ter training times. The performance achieved for the NER and sentiment analysis tasks ranged from -0.8% to -6.2%. This indicates a trade-off between the use of training parameters and the resulting performance. However, the Prefix-Tuning and UniPELT methods failed to provide consistent results on the summarization task.