Text backdoor detection using an interpretable RNN abstract model
Deep neural networks (DNNs) are known to be inherently vulnerable to malicious attacks such as the adversarial attack and the backdoor attack. The former is crafted by adding small perturbations to benign inputs so as to fool a DNN. The latter generally embeds a hidden pattern in a DNN by poisoning...
Saved in:
Main Authors: | FAN, Ming, SI, Ziliang, XIE, Xiaofei, LIU, Yang, LIU, Ting |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2021
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/7118 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
Similar Items
-
Towards interpreting recurrent neural networks through probabilistic abstraction
by: DONG, Guoliang, et al.
Published: (2020) -
Stealthy backdoor attack for code models
by: YANG, Zhou, et al.
Published: (2024) -
Evaluation of backdoor attacks and defenses to deep neural networks
by: Ooi, Ying Xuan
Published: (2024) -
DeepStellar: Model-based quantitative analysis of stateful deep learning systems
by: DU, Xiaoning, et al.
Published: (2019) -
BADFL: Backdoor attack defense in federated learning from local model perspective
by: ZHANG, Haiyan, et al.
Published: (2024)