Privacy and security issues in deep learning : a survey

Deep Learning (DL) algorithms based on artificial neural networks have achieved remarkable success and are being extensively applied in a variety of application domains, ranging from image classification, automatic driving, natural language processing to medical diagnosis, credit risk assessment, in...

Full description

Saved in:
Bibliographic Details
Main Authors: Liu, Ximeng, Xie, Lehui, Wang, Yaopeng, Zou, Jian, Xiong, Jinbo, Ying, Zuobin, Vasilakos, Athanasios V.
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2021
Subjects:
Online Access:https://hdl.handle.net/10356/145999
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-145999
record_format dspace
spelling sg-ntu-dr.10356-1459992021-01-20T03:16:03Z Privacy and security issues in deep learning : a survey Liu, Ximeng Xie, Lehui Wang, Yaopeng Zou, Jian Xiong, Jinbo Ying, Zuobin Vasilakos, Athanasios V. School of Electrical and Electronic Engineering Engineering::Electrical and electronic engineering Deep Learning DL Privacy Deep Learning (DL) algorithms based on artificial neural networks have achieved remarkable success and are being extensively applied in a variety of application domains, ranging from image classification, automatic driving, natural language processing to medical diagnosis, credit risk assessment, intrusion detection. However, the privacy and security issues of DL have been revealed that the DL model can be stolen or reverse engineered, sensitive training data can be inferred, even a recognizable face image of the victim can be recovered. Besides, the recent works have found that the DL model is vulnerable to adversarial examples perturbed by imperceptible noised, which can lead the DL model to predict wrongly with high confidence. In this paper, we first briefly introduces the four types of attacks and privacy-preserving techniques in DL. We then review and summarize the attack and defense methods associated with DL privacy and security in recent years. To demonstrate that security threats really exist in the real world, we also reviewed the adversarial attacks under the physical condition. Finally, we discuss current challenges and open problems regarding privacy and security issues in DL. Published version 2021-01-20T03:16:03Z 2021-01-20T03:16:03Z 2020 Journal Article Liu, X., Xie, L., Wang, Y., Zou, J., Xiong, J., Ying, Z., & Vasilakos, A. V. (2021). Privacy and security issues in deep learning : a survey. IEEE Access, 9, 4566-4593. doi:10.1109/ACCESS.2020.3045078 2169-3536 https://hdl.handle.net/10356/145999 10.1109/ACCESS.2020.3045078 2-s2.0-85098748130 9 4566 4593 en IEEE Access © 2020 IEEE. This journal is 100% open access, which means that all content is freely available without charge to users or their institutions. All articles accepted after 12 June 2019 are published under a CC BY 4.0 license, and the author retains copyright. Users are allowed to read, download, copy, distribute, print, search, or link to the full texts of the articles, or use them for any other lawful purpose, as long as proper attribution is given. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
Deep Learning
DL Privacy
spellingShingle Engineering::Electrical and electronic engineering
Deep Learning
DL Privacy
Liu, Ximeng
Xie, Lehui
Wang, Yaopeng
Zou, Jian
Xiong, Jinbo
Ying, Zuobin
Vasilakos, Athanasios V.
Privacy and security issues in deep learning : a survey
description Deep Learning (DL) algorithms based on artificial neural networks have achieved remarkable success and are being extensively applied in a variety of application domains, ranging from image classification, automatic driving, natural language processing to medical diagnosis, credit risk assessment, intrusion detection. However, the privacy and security issues of DL have been revealed that the DL model can be stolen or reverse engineered, sensitive training data can be inferred, even a recognizable face image of the victim can be recovered. Besides, the recent works have found that the DL model is vulnerable to adversarial examples perturbed by imperceptible noised, which can lead the DL model to predict wrongly with high confidence. In this paper, we first briefly introduces the four types of attacks and privacy-preserving techniques in DL. We then review and summarize the attack and defense methods associated with DL privacy and security in recent years. To demonstrate that security threats really exist in the real world, we also reviewed the adversarial attacks under the physical condition. Finally, we discuss current challenges and open problems regarding privacy and security issues in DL.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Liu, Ximeng
Xie, Lehui
Wang, Yaopeng
Zou, Jian
Xiong, Jinbo
Ying, Zuobin
Vasilakos, Athanasios V.
format Article
author Liu, Ximeng
Xie, Lehui
Wang, Yaopeng
Zou, Jian
Xiong, Jinbo
Ying, Zuobin
Vasilakos, Athanasios V.
author_sort Liu, Ximeng
title Privacy and security issues in deep learning : a survey
title_short Privacy and security issues in deep learning : a survey
title_full Privacy and security issues in deep learning : a survey
title_fullStr Privacy and security issues in deep learning : a survey
title_full_unstemmed Privacy and security issues in deep learning : a survey
title_sort privacy and security issues in deep learning : a survey
publishDate 2021
url https://hdl.handle.net/10356/145999
_version_ 1690658298186956800