Breaking neural reasoning architectures with metamorphic relation-based adversarial examples

The ability to read, reason, and infer lies at the heart of neural reasoning architectures. After all, the ability to perform logical reasoning over language remains a coveted goal of Artificial Intelligence. To this end, models such as the Turing-complete differentiable neural computer (DNC) boast...

Full description

Saved in:
Bibliographic Details
Main Authors: CHAN, Alvin, MA, Lei, JUEFEI-XU, Felix, ONG, Yew-Soon, XIE, Xiaofei, XUE, Minhui, LIU, Yang
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2021
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7050
https://ink.library.smu.edu.sg/context/sis_research/article/8053/viewcontent/felix_tnnls21_dnc.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8053
record_format dspace
spelling sg-smu-ink.sis_research-80532022-04-07T09:08:28Z Breaking neural reasoning architectures with metamorphic relation-based adversarial examples CHAN, Alvin MA, Lei JUEFEI-XU, Felix ONG, Yew-Soon XIE, Xiaofei XUE, Minhui LIU, Yang The ability to read, reason, and infer lies at the heart of neural reasoning architectures. After all, the ability to perform logical reasoning over language remains a coveted goal of Artificial Intelligence. To this end, models such as the Turing-complete differentiable neural computer (DNC) boast of real logical reasoning capabilities, along with the ability to reason beyond simple surface-level matching. In this brief, we propose the first probe into DNC's logical reasoning capabilities with a focus on text-based question answering (QA). More concretely, we propose a conceptually simple but effective adversarial attack based on metamorphic relations. Our proposed adversarial attack reduces DNCs' state-of-the-art accuracy from 100% to 1.5% in the worst case, exposing weaknesses and susceptibilities in modern neural reasoning architectures. We further empirically explore possibilities to defend against such attacks and demonstrate the utility of our adversarial framework as a simple scalable method to improve model adversarial robustness. 2021-04-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7050 info:doi/10.1109/TNNLS.2021.3072166 https://ink.library.smu.edu.sg/context/sis_research/article/8053/viewcontent/felix_tnnls21_dnc.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Task analysism Cognition Plugs;Perturbation methods Memory modules Computer architecture Computational modeling Adversarial examples deep learning differentiable neural computer (DNC) supervised learning OS and Networks Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Task analysism
Cognition
Plugs;Perturbation methods
Memory modules
Computer architecture
Computational modeling
Adversarial examples
deep learning
differentiable neural computer (DNC)
supervised learning
OS and Networks
Software Engineering
spellingShingle Task analysism
Cognition
Plugs;Perturbation methods
Memory modules
Computer architecture
Computational modeling
Adversarial examples
deep learning
differentiable neural computer (DNC)
supervised learning
OS and Networks
Software Engineering
CHAN, Alvin
MA, Lei
JUEFEI-XU, Felix
ONG, Yew-Soon
XIE, Xiaofei
XUE, Minhui
LIU, Yang
Breaking neural reasoning architectures with metamorphic relation-based adversarial examples
description The ability to read, reason, and infer lies at the heart of neural reasoning architectures. After all, the ability to perform logical reasoning over language remains a coveted goal of Artificial Intelligence. To this end, models such as the Turing-complete differentiable neural computer (DNC) boast of real logical reasoning capabilities, along with the ability to reason beyond simple surface-level matching. In this brief, we propose the first probe into DNC's logical reasoning capabilities with a focus on text-based question answering (QA). More concretely, we propose a conceptually simple but effective adversarial attack based on metamorphic relations. Our proposed adversarial attack reduces DNCs' state-of-the-art accuracy from 100% to 1.5% in the worst case, exposing weaknesses and susceptibilities in modern neural reasoning architectures. We further empirically explore possibilities to defend against such attacks and demonstrate the utility of our adversarial framework as a simple scalable method to improve model adversarial robustness.
format text
author CHAN, Alvin
MA, Lei
JUEFEI-XU, Felix
ONG, Yew-Soon
XIE, Xiaofei
XUE, Minhui
LIU, Yang
author_facet CHAN, Alvin
MA, Lei
JUEFEI-XU, Felix
ONG, Yew-Soon
XIE, Xiaofei
XUE, Minhui
LIU, Yang
author_sort CHAN, Alvin
title Breaking neural reasoning architectures with metamorphic relation-based adversarial examples
title_short Breaking neural reasoning architectures with metamorphic relation-based adversarial examples
title_full Breaking neural reasoning architectures with metamorphic relation-based adversarial examples
title_fullStr Breaking neural reasoning architectures with metamorphic relation-based adversarial examples
title_full_unstemmed Breaking neural reasoning architectures with metamorphic relation-based adversarial examples
title_sort breaking neural reasoning architectures with metamorphic relation-based adversarial examples
publisher Institutional Knowledge at Singapore Management University
publishDate 2021
url https://ink.library.smu.edu.sg/sis_research/7050
https://ink.library.smu.edu.sg/context/sis_research/article/8053/viewcontent/felix_tnnls21_dnc.pdf
_version_ 1770576195141763072