Text backdoor detection using an interpretable RNN abstract model
Deep neural networks (DNNs) are known to be inherently vulnerable to malicious attacks such as the adversarial attack and the backdoor attack. The former is crafted by adding small perturbations to benign inputs so as to fool a DNN. The latter generally embeds a hidden pattern in a DNN by poisoning...
Saved in:
Main Authors: | FAN, Ming, SI, Ziliang, XIE, Xiaofei, LIU, Yang, LIU, Ting |
---|---|
格式: | text |
語言: | English |
出版: |
Institutional Knowledge at Singapore Management University
2021
|
主題: | |
在線閱讀: | https://ink.library.smu.edu.sg/sis_research/7118 |
標簽: |
添加標簽
沒有標簽, 成為第一個標記此記錄!
|
機構: | Singapore Management University |
語言: | English |
相似書籍
-
Towards interpreting recurrent neural networks through probabilistic abstraction
由: DONG, Guoliang, et al.
出版: (2020) -
Stealthy backdoor attack for code models
由: YANG, Zhou, et al.
出版: (2024) -
DeepStellar: Model-based quantitative analysis of stateful deep learning systems
由: DU, Xiaoning, et al.
出版: (2019) -
Evaluation of backdoor attacks and defenses to deep neural networks
由: Ooi, Ying Xuan
出版: (2024) -
Deepcause: Verifying neural networks with abstraction refinement
由: NGUYEN HUA GIA PHUC,
出版: (2022)