Adversarial training using meta-learning for BERT
Deep learning is currently the most successful method of semantic analysis in natural language processing. However, in recent years, many variants of carefully crafted inputs designed to cause misclassification, known as adversarial attacks, have been engineered with tremendous success. One well-...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/156635 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-156635 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1566352022-04-21T07:08:25Z Adversarial training using meta-learning for BERT Low, Timothy Jing Haen Joty Shafiq Rayhan School of Computer Science and Engineering srjoty@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Deep learning is currently the most successful method of semantic analysis in natural language processing. However, in recent years, many variants of carefully crafted inputs designed to cause misclassification, known as adversarial attacks, have been engineered with tremendous success. One well-known, efficient method to develop models to be robust against adversarial attacks is known as adversarial training, where models are iteratively trained on samples produces by the specific attack algorithm. However, adversarial training only works when the model has access to the attack generation algorithm or a large dataset of attack samples, and so cannot defend against attacks of which they have access to a low number of samples. This project proposes to overcome this challenge using meta-learning, which uses a large number of similar tasks from a different domain to train a classifier to learn another task for which a small number of labelled samples are available. We show that by using the Model-Agnostic Meta-Learning algorithm in adversarial training, a model trained on a large number of different adversarial attacks can become more robust to an adversarial attack that it has few samples of. This project will also explore augmenting the training set with a large number of non-adversarial perturbations, in order to possibly better mitigate adversarial attacks Bachelor of Engineering (Computer Science) 2022-04-21T07:08:25Z 2022-04-21T07:08:25Z 2022 Final Year Project (FYP) Low, T. J. H. (2022). Adversarial training using meta-learning for BERT. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/156635 https://hdl.handle.net/10356/156635 en SCSE21-0524 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence |
spellingShingle |
Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Low, Timothy Jing Haen Adversarial training using meta-learning for BERT |
description |
Deep learning is currently the most successful method of semantic analysis in natural language
processing. However, in recent years, many variants of carefully crafted inputs designed to cause
misclassification, known as adversarial attacks, have been engineered with tremendous success. One
well-known, efficient method to develop models to be robust against adversarial attacks is known as
adversarial training, where models are iteratively trained on samples produces by the specific attack
algorithm. However, adversarial training only works when the model has access to the attack
generation algorithm or a large dataset of attack samples, and so cannot defend against attacks of
which they have access to a low number of samples. This project proposes to overcome this challenge
using meta-learning, which uses a large number of similar tasks from a different domain to train a
classifier to learn another task for which a small number of labelled samples are available. We show
that by using the Model-Agnostic Meta-Learning algorithm in adversarial training, a model trained on
a large number of different adversarial attacks can become more robust to an adversarial attack that it
has few samples of. This project will also explore augmenting the training set with a large number of
non-adversarial perturbations, in order to possibly better mitigate adversarial attacks |
author2 |
Joty Shafiq Rayhan |
author_facet |
Joty Shafiq Rayhan Low, Timothy Jing Haen |
format |
Final Year Project |
author |
Low, Timothy Jing Haen |
author_sort |
Low, Timothy Jing Haen |
title |
Adversarial training using meta-learning for BERT |
title_short |
Adversarial training using meta-learning for BERT |
title_full |
Adversarial training using meta-learning for BERT |
title_fullStr |
Adversarial training using meta-learning for BERT |
title_full_unstemmed |
Adversarial training using meta-learning for BERT |
title_sort |
adversarial training using meta-learning for bert |
publisher |
Nanyang Technological University |
publishDate |
2022 |
url |
https://hdl.handle.net/10356/156635 |
_version_ |
1731235790023819264 |