An analysis of adversarial algorithm techniques in image recognition and their countermeasures

The ability of neural network models to generalise and identify unseen data allows for neural networks to operate outside of what it has been trained on, but makes it vulnerable to data samples altered in human imperceptible ways to produce incorrect predictions. This project aims to experimentally...

Full description

Saved in:
Bibliographic Details
Main Author: Tan, Alastair Song Xin
Other Authors: Kong Wai-Kin Adams
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2021
Subjects:
Online Access:https://hdl.handle.net/10356/153433
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:The ability of neural network models to generalise and identify unseen data allows for neural networks to operate outside of what it has been trained on, but makes it vulnerable to data samples altered in human imperceptible ways to produce incorrect predictions. This project aims to experimentally test some adversarial algorithms used to fool neural networks, and examine some defensive techniques used to mitigate or prevent such attacks. The MNIST digit dataset, Tensorflow and the Cleverhans Library were used to collect the results required, and it was identified that dropping out neurons and adversarial training not only provided some level of protection against basic adversarial attacks, but improved a model’s capability to generalise and identify unseen, non-adversarial samples.