Privacy-preserving federated deep learning with irregular users

Federated deep learning has been widely used in various fields. To protect data privacy, many privacy-preserving approaches have also been designed and implemented in various scenarios. However, existing works rarely consider a fundamental issue that the data shared by certain users (called irregula...

Full description

Saved in:
Bibliographic Details
Main Authors: XU, Guowen, LI, Hongwei, ZHANG, Yun, XU, Shengmin, NING, Jianting, DENG, Robert H.
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2021
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/5181
https://ink.library.smu.edu.sg/context/sis_research/article/6184/viewcontent/Privacy_Preserving_Federated_Deep_Learning_Irregular_Users_2020_av.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-6184
record_format dspace
spelling sg-smu-ink.sis_research-61842022-04-18T10:37:38Z Privacy-preserving federated deep learning with irregular users XU, Guowen LI, Hongwei ZHANG, Yun XU, Shengmin NING, Jianting DENG, Robert H. Federated deep learning has been widely used in various fields. To protect data privacy, many privacy-preserving approaches have also been designed and implemented in various scenarios. However, existing works rarely consider a fundamental issue that the data shared by certain users (called irregular users) may be of low quality. Obviously, in a federated training process, data shared by many irregular users may impair the training accuracy, or worse, lead to the uselessness of the final model. In this paper, we propose PPFDL, a Privacy-Preserving Federated Deep Learning framework with irregular users. In specific, we design a novel solution to reduce the negative impact of irregular users on the training accuracy, which guarantees that the training results are mainly calculated from the contribution of high-quality data. Meanwhile, we exploit Yaos garbled circuits and additively homomorphic cryptosystems to ensure the confidentiality of all user-related information. Moreover, PPFDL is also robustness to users dropping out during the whole implementation. This means that each user can be offline at any subprocess of training, as long as the remaining online users can still complete the training task. Extensive experiments demonstrate the superior performance of PPFDL in terms of training accuracy, computation, and communication overheads. 2021-03-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/5181 info:doi/10.1109/TDSC.2020.3005909 https://ink.library.smu.edu.sg/context/sis_research/article/6184/viewcontent/Privacy_Preserving_Federated_Deep_Learning_Irregular_Users_2020_av.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Collaborative learning deep learning privacy Information Security Numerical Analysis and Scientific Computing
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Collaborative learning
deep learning
privacy
Information Security
Numerical Analysis and Scientific Computing
spellingShingle Collaborative learning
deep learning
privacy
Information Security
Numerical Analysis and Scientific Computing
XU, Guowen
LI, Hongwei
ZHANG, Yun
XU, Shengmin
NING, Jianting
DENG, Robert H.
Privacy-preserving federated deep learning with irregular users
description Federated deep learning has been widely used in various fields. To protect data privacy, many privacy-preserving approaches have also been designed and implemented in various scenarios. However, existing works rarely consider a fundamental issue that the data shared by certain users (called irregular users) may be of low quality. Obviously, in a federated training process, data shared by many irregular users may impair the training accuracy, or worse, lead to the uselessness of the final model. In this paper, we propose PPFDL, a Privacy-Preserving Federated Deep Learning framework with irregular users. In specific, we design a novel solution to reduce the negative impact of irregular users on the training accuracy, which guarantees that the training results are mainly calculated from the contribution of high-quality data. Meanwhile, we exploit Yaos garbled circuits and additively homomorphic cryptosystems to ensure the confidentiality of all user-related information. Moreover, PPFDL is also robustness to users dropping out during the whole implementation. This means that each user can be offline at any subprocess of training, as long as the remaining online users can still complete the training task. Extensive experiments demonstrate the superior performance of PPFDL in terms of training accuracy, computation, and communication overheads.
format text
author XU, Guowen
LI, Hongwei
ZHANG, Yun
XU, Shengmin
NING, Jianting
DENG, Robert H.
author_facet XU, Guowen
LI, Hongwei
ZHANG, Yun
XU, Shengmin
NING, Jianting
DENG, Robert H.
author_sort XU, Guowen
title Privacy-preserving federated deep learning with irregular users
title_short Privacy-preserving federated deep learning with irregular users
title_full Privacy-preserving federated deep learning with irregular users
title_fullStr Privacy-preserving federated deep learning with irregular users
title_full_unstemmed Privacy-preserving federated deep learning with irregular users
title_sort privacy-preserving federated deep learning with irregular users
publisher Institutional Knowledge at Singapore Management University
publishDate 2021
url https://ink.library.smu.edu.sg/sis_research/5181
https://ink.library.smu.edu.sg/context/sis_research/article/6184/viewcontent/Privacy_Preserving_Federated_Deep_Learning_Irregular_Users_2020_av.pdf
_version_ 1770575322748551168