ADVERSARIAL ATTACK ON NEURAL MACHINE TRANSLATION MODELS AS DISCRETE OPTIMIZATION
Machine Translation models are commonly trained only on perfect grammar corpora. Training on perfect grammar corpora might predispose the model to be biased against minorities from non-standard linguistics backgrounds. Adversarial attack is one way to protect machine learning models from this bia...
Saved in:
Main Author: | |
---|---|
Format: | Final Project |
Language: | Indonesia |
Online Access: | https://digilib.itb.ac.id/gdl/view/65291 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Institut Teknologi Bandung |
Language: | Indonesia |
id |
id-itb.:65291 |
---|---|
spelling |
id-itb.:652912022-06-22T08:55:02ZADVERSARIAL ATTACK ON NEURAL MACHINE TRANSLATION MODELS AS DISCRETE OPTIMIZATION Kuwanto, Garry Indonesia Final Project Machine Translation, Natural Language Processing, Adversarial Attack, Discrete Optimization, Genetic Algorithm, Simulated Annealing, Meta-heuristics. INSTITUT TEKNOLOGI BANDUNG https://digilib.itb.ac.id/gdl/view/65291 Machine Translation models are commonly trained only on perfect grammar corpora. Training on perfect grammar corpora might predispose the model to be biased against minorities from non-standard linguistics backgrounds. Adversarial attack is one way to protect machine learning models from this bias. In this final project, We perform an adversarial attack by means of perturbation of words using the different morphological forms. This perturbation crafts adversarial examples with non-perfect grammar for English sentences. Because of how choosing the morphological forms is seen as finding an optimal choice from a discrete search space, we can see this as a discrete optimization problem. Here, we focus on exploring three different discrete optimization method, Genetic Algorithm, Simulated Annealing, and a modified version of a reinforcement learning algorithm, the REINFORCE algorithm. Comparison with the greedy algorithm as a commonly used method, our results show little difference between sentence crafted from these methods. In terms of efficiency, Simulated Annealing is a significantly faster algorithm, three times faster than the slowest algorithm (Genetic Algorithm) and faster than the second best algorithm (REINFORCE) by a factor of two. text |
institution |
Institut Teknologi Bandung |
building |
Institut Teknologi Bandung Library |
continent |
Asia |
country |
Indonesia Indonesia |
content_provider |
Institut Teknologi Bandung |
collection |
Digital ITB |
language |
Indonesia |
description |
Machine Translation models are commonly trained only on perfect grammar
corpora. Training on perfect grammar corpora might predispose the model to be
biased against minorities from non-standard linguistics backgrounds. Adversarial
attack is one way to protect machine learning models from this bias.
In this final project, We perform an adversarial attack by means of perturbation
of words using the different morphological forms. This perturbation crafts adversarial
examples with non-perfect grammar for English sentences. Because of how
choosing the morphological forms is seen as finding an optimal choice from a
discrete search space, we can see this as a discrete optimization problem. Here, we
focus on exploring three different discrete optimization method, Genetic Algorithm,
Simulated Annealing, and a modified version of a reinforcement learning algorithm,
the REINFORCE algorithm. Comparison with the greedy algorithm as a commonly
used method, our results show little difference between sentence crafted from these
methods. In terms of efficiency, Simulated Annealing is a significantly faster
algorithm, three times faster than the slowest algorithm (Genetic Algorithm) and
faster than the second best algorithm (REINFORCE) by a factor of two. |
format |
Final Project |
author |
Kuwanto, Garry |
spellingShingle |
Kuwanto, Garry ADVERSARIAL ATTACK ON NEURAL MACHINE TRANSLATION MODELS AS DISCRETE OPTIMIZATION |
author_facet |
Kuwanto, Garry |
author_sort |
Kuwanto, Garry |
title |
ADVERSARIAL ATTACK ON NEURAL MACHINE TRANSLATION MODELS AS DISCRETE OPTIMIZATION |
title_short |
ADVERSARIAL ATTACK ON NEURAL MACHINE TRANSLATION MODELS AS DISCRETE OPTIMIZATION |
title_full |
ADVERSARIAL ATTACK ON NEURAL MACHINE TRANSLATION MODELS AS DISCRETE OPTIMIZATION |
title_fullStr |
ADVERSARIAL ATTACK ON NEURAL MACHINE TRANSLATION MODELS AS DISCRETE OPTIMIZATION |
title_full_unstemmed |
ADVERSARIAL ATTACK ON NEURAL MACHINE TRANSLATION MODELS AS DISCRETE OPTIMIZATION |
title_sort |
adversarial attack on neural machine translation models as discrete optimization |
url |
https://digilib.itb.ac.id/gdl/view/65291 |
_version_ |
1822932699031011328 |