ADVERSARIAL ROBUSTNES TESTING ON PRETRAINED LANGUAGE MODEL IN TEXT CLASSIFICATION
Text representation technology is growing with the existence of monolingual and multilingual pre-trained language models. Language models are increasingly being used, especially on text classification problems. One phenomenon that occur in textual data is code-mixing and synonym replacement. With...
Saved in:
Main Author: | |
---|---|
Format: | Final Project |
Language: | Indonesia |
Online Access: | https://digilib.itb.ac.id/gdl/view/65818 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Institut Teknologi Bandung |
Language: | Indonesia |
Summary: | Text representation technology is growing with the existence of monolingual and
multilingual pre-trained language models. Language models are increasingly being
used, especially on text classification problems. One phenomenon that occur in
textual data is code-mixing and synonym replacement. With the increasingly vital
role of language models training, it is necessary to do further evaluation whether the
existing pre-trained language models are good enough to handle various cases on
textual communication. One technique that can be used is adversarial attack.
Adversarial attack (Jin et al., 2020) has the ability to find words that most contribute
to the label prediction by a model (vulnerable words). By using adversarial attack
technique, these vulnerable words will be translated to simulate the phenomenon
of code-mixing and synonym replacement perturbation. The perturbed text will be
evaluated with a semantic similarity score to preserve its semantic meaning.
Experiments were carried out with two text classification tasks and the results showed
that all language models experienced a decrease in performance. In the case of codemixing Indonesian with foreign languages that are not related to Indonesian, the
XLM-R model outperforms the IndoBERT model, while in the case of code-mixing
languages related to Indonesian, the IndoBERT model outperforms the XLM-R
model. Experimental results also show that increasing the size of the model increases
the robustness of the model.
|
---|