Adversarial training using meta-learning for BERT

Deep learning is currently the most successful method of semantic analysis in natural language processing. However, in recent years, many variants of carefully crafted inputs designed to cause misclassification, known as adversarial attacks, have been engineered with tremendous success. One well-...

Full description

Saved in:
Bibliographic Details
Main Author: Low, Timothy Jing Haen
Other Authors: Joty Shafiq Rayhan
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2022
Subjects:
Online Access:https://hdl.handle.net/10356/156635
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-156635
record_format dspace
spelling sg-ntu-dr.10356-1566352022-04-21T07:08:25Z Adversarial training using meta-learning for BERT Low, Timothy Jing Haen Joty Shafiq Rayhan School of Computer Science and Engineering srjoty@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Deep learning is currently the most successful method of semantic analysis in natural language processing. However, in recent years, many variants of carefully crafted inputs designed to cause misclassification, known as adversarial attacks, have been engineered with tremendous success. One well-known, efficient method to develop models to be robust against adversarial attacks is known as adversarial training, where models are iteratively trained on samples produces by the specific attack algorithm. However, adversarial training only works when the model has access to the attack generation algorithm or a large dataset of attack samples, and so cannot defend against attacks of which they have access to a low number of samples. This project proposes to overcome this challenge using meta-learning, which uses a large number of similar tasks from a different domain to train a classifier to learn another task for which a small number of labelled samples are available. We show that by using the Model-Agnostic Meta-Learning algorithm in adversarial training, a model trained on a large number of different adversarial attacks can become more robust to an adversarial attack that it has few samples of. This project will also explore augmenting the training set with a large number of non-adversarial perturbations, in order to possibly better mitigate adversarial attacks Bachelor of Engineering (Computer Science) 2022-04-21T07:08:25Z 2022-04-21T07:08:25Z 2022 Final Year Project (FYP) Low, T. J. H. (2022). Adversarial training using meta-learning for BERT. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/156635 https://hdl.handle.net/10356/156635 en SCSE21-0524 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence
spellingShingle Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence
Low, Timothy Jing Haen
Adversarial training using meta-learning for BERT
description Deep learning is currently the most successful method of semantic analysis in natural language processing. However, in recent years, many variants of carefully crafted inputs designed to cause misclassification, known as adversarial attacks, have been engineered with tremendous success. One well-known, efficient method to develop models to be robust against adversarial attacks is known as adversarial training, where models are iteratively trained on samples produces by the specific attack algorithm. However, adversarial training only works when the model has access to the attack generation algorithm or a large dataset of attack samples, and so cannot defend against attacks of which they have access to a low number of samples. This project proposes to overcome this challenge using meta-learning, which uses a large number of similar tasks from a different domain to train a classifier to learn another task for which a small number of labelled samples are available. We show that by using the Model-Agnostic Meta-Learning algorithm in adversarial training, a model trained on a large number of different adversarial attacks can become more robust to an adversarial attack that it has few samples of. This project will also explore augmenting the training set with a large number of non-adversarial perturbations, in order to possibly better mitigate adversarial attacks
author2 Joty Shafiq Rayhan
author_facet Joty Shafiq Rayhan
Low, Timothy Jing Haen
format Final Year Project
author Low, Timothy Jing Haen
author_sort Low, Timothy Jing Haen
title Adversarial training using meta-learning for BERT
title_short Adversarial training using meta-learning for BERT
title_full Adversarial training using meta-learning for BERT
title_fullStr Adversarial training using meta-learning for BERT
title_full_unstemmed Adversarial training using meta-learning for BERT
title_sort adversarial training using meta-learning for bert
publisher Nanyang Technological University
publishDate 2022
url https://hdl.handle.net/10356/156635
_version_ 1731235790023819264