Efficient dropout-resilient aggregation for privacy-preserving machine learning

Machine learning (ML) has been widely recognized as an enabler of the global trend of digital transformation. With the increasing adoption of data-hungry machine learning algorithms, personal data privacy has emerged as one of the key concerns that could hinder the success of digital transformation....

Full description

Saved in:
Bibliographic Details
Main Authors: Liu, Ziyao, Guo, Jiale, Lam, Kwok-Yan, Zhao, Jun
Other Authors: School of Computer Science and Engineering
Format: Article
Language:English
Published: 2022
Subjects:
Online Access:https://hdl.handle.net/10356/162985
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-162985
record_format dspace
spelling sg-ntu-dr.10356-1629852023-05-19T15:36:26Z Efficient dropout-resilient aggregation for privacy-preserving machine learning Liu, Ziyao Guo, Jiale Lam, Kwok-Yan Zhao, Jun School of Computer Science and Engineering Engineering::Computer science and engineering Secure Aggregation Privacy-Preserving Machine Learning Machine learning (ML) has been widely recognized as an enabler of the global trend of digital transformation. With the increasing adoption of data-hungry machine learning algorithms, personal data privacy has emerged as one of the key concerns that could hinder the success of digital transformation. As such, Privacy-Preserving Machine Learning (PPML) has received much attention of the machine learning community, from academic researchers to industry practitioners to government regulators. However, organizations are faced with the dilemma that, on the one hand, they are encouraged to share data to enhance ML performance, but on the other hand, they could potentially be breaching the relevant data privacy regulations. Practical PPML typically allows multiple participants to individually train their ML models, which are then aggregated to construct a global model in a privacy-preserving manner, e.g., based on multi-party computation or homomorphic encryption. Nevertheless, in most important applications of large-scale PPML, e.g., by aggregating clients’ gradients to update a global model for federated learning, such as consumer behavior modeling of mobile application services, some participants are inevitably resource-constrained mobile devices, which may drop out of the PPML system due to their mobility nature [1]. Therefore, the resilience of privacy-preserving aggregation has become an important problem to be tackled because of its real-world application potential and impacts. In this paper, we propose a scalable privacy-preserving aggregation scheme that can tolerate dropout by participants at any time, and is secure against both semi-honest and active malicious adversaries by setting proper system parameters. By replacing communication-intensive building blocks with a seed homomorphic pseudo-random generator, and relying on the additive homomorphic property of Shamir secret sharing scheme, our scheme outperforms state-of-the-art schemes by up to 6.37× in runtime and provides a stronger dropout-resilience. The simplicity of our scheme makes it attractive both for implementation and for further improvements. National Research Foundation (NRF) Submitted/Accepted version This research is supported by the National Research Foundation, Singapore under its Strategic Capability Research Centres Funding Initiative. 2022-11-14T07:38:04Z 2022-11-14T07:38:04Z 2022 Journal Article Liu, Z., Guo, J., Lam, K. & Zhao, J. (2022). Efficient dropout-resilient aggregation for privacy-preserving machine learning. IEEE Transactions On Information Forensics and Security, 14(8), 3163592-. https://dx.doi.org/10.1109/TIFS.2022.3163592 1556-6013 https://hdl.handle.net/10356/162985 10.1109/TIFS.2022.3163592 2-s2.0-85128597543 8 14 3163592 en IEEE Transactions on Information Forensics and Security © 2021 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. The published version is available at: https://doi.org/10.1109/TIFS.2022.3163592. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering
Secure Aggregation
Privacy-Preserving Machine Learning
spellingShingle Engineering::Computer science and engineering
Secure Aggregation
Privacy-Preserving Machine Learning
Liu, Ziyao
Guo, Jiale
Lam, Kwok-Yan
Zhao, Jun
Efficient dropout-resilient aggregation for privacy-preserving machine learning
description Machine learning (ML) has been widely recognized as an enabler of the global trend of digital transformation. With the increasing adoption of data-hungry machine learning algorithms, personal data privacy has emerged as one of the key concerns that could hinder the success of digital transformation. As such, Privacy-Preserving Machine Learning (PPML) has received much attention of the machine learning community, from academic researchers to industry practitioners to government regulators. However, organizations are faced with the dilemma that, on the one hand, they are encouraged to share data to enhance ML performance, but on the other hand, they could potentially be breaching the relevant data privacy regulations. Practical PPML typically allows multiple participants to individually train their ML models, which are then aggregated to construct a global model in a privacy-preserving manner, e.g., based on multi-party computation or homomorphic encryption. Nevertheless, in most important applications of large-scale PPML, e.g., by aggregating clients’ gradients to update a global model for federated learning, such as consumer behavior modeling of mobile application services, some participants are inevitably resource-constrained mobile devices, which may drop out of the PPML system due to their mobility nature [1]. Therefore, the resilience of privacy-preserving aggregation has become an important problem to be tackled because of its real-world application potential and impacts. In this paper, we propose a scalable privacy-preserving aggregation scheme that can tolerate dropout by participants at any time, and is secure against both semi-honest and active malicious adversaries by setting proper system parameters. By replacing communication-intensive building blocks with a seed homomorphic pseudo-random generator, and relying on the additive homomorphic property of Shamir secret sharing scheme, our scheme outperforms state-of-the-art schemes by up to 6.37× in runtime and provides a stronger dropout-resilience. The simplicity of our scheme makes it attractive both for implementation and for further improvements.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Liu, Ziyao
Guo, Jiale
Lam, Kwok-Yan
Zhao, Jun
format Article
author Liu, Ziyao
Guo, Jiale
Lam, Kwok-Yan
Zhao, Jun
author_sort Liu, Ziyao
title Efficient dropout-resilient aggregation for privacy-preserving machine learning
title_short Efficient dropout-resilient aggregation for privacy-preserving machine learning
title_full Efficient dropout-resilient aggregation for privacy-preserving machine learning
title_fullStr Efficient dropout-resilient aggregation for privacy-preserving machine learning
title_full_unstemmed Efficient dropout-resilient aggregation for privacy-preserving machine learning
title_sort efficient dropout-resilient aggregation for privacy-preserving machine learning
publishDate 2022
url https://hdl.handle.net/10356/162985
_version_ 1772826019544170496