Byzantine-resilient decentralized stochastic gradient descent
Decentralized learning has gained great popularity to improve learning efficiency and preserve data privacy. Each computing node makes equal contribution to collaboratively learn a Deep Learning model. The elimination of centralized Parameter Servers (PS) can effectively address many issues such as...
Saved in:
Main Authors: | , , , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/179057 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-179057 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1790572024-07-18T00:32:00Z Byzantine-resilient decentralized stochastic gradient descent Guo, Shangwei Zhang, Tianwei Yu, Han Xie, Xiaofei Ma, Lei Xiang, Tao Liu, Yang College of Computing and Data Science School of Computer Science and Engineering Computer and Information Science Artificial intelligence Federated learning Decentralized learning has gained great popularity to improve learning efficiency and preserve data privacy. Each computing node makes equal contribution to collaboratively learn a Deep Learning model. The elimination of centralized Parameter Servers (PS) can effectively address many issues such as privacy, performance bottleneck and single-point-failure. However, how to achieve Byzantine Fault Tolerance in decentralized learning systems is rarely explored, although this problem has been extensively studied in centralized systems. In this paper, we present an in-depth study towards the Byzantine resilience of decentralized learning systems with two contributions. First, from the adversarial perspective, we theoretically illustrate that Byzantine attacks are more dangerous and feasible in decentralized learning systems: even one malicious participant can arbitrarily alter the models of other participants by sending carefully crafted updates to its neighbors. Second, from the defense perspective, we propose Ubar, a novel algorithm to enhance decentralized learning with Byzantine Fault Tolerance. Specifically, Ubar provides a U niform B yzantine-resilient A ggregation R ule for benign nodes to select the useful parameter updates and filter out the malicious ones in each training iteration. It guarantees that each benign node in a decentralized system can train a correct model under very strong Byzantine attacks with an arbitrary number of faulty nodes. We conduct extensive experiments on standard image classification tasks and the results indicate that Ubar can effectively defeat both simple and sophisticated Byzantine attacks with higher performance efficiency than existing solutions. Agency for Science, Technology and Research (A*STAR) AI Singapore Ministry of Education (MOE) Nanyang Technological University National Research Foundation (NRF) Submitted/Accepted version This work was supported in part by the National Natural Science Foundation of China under Grant 62102052; in part by Singapore Ministry of Education Academic Research Fund Tier 1 under Award RS02/19 and Award 2018-T1-002-069; in part by the National Research Foundation, Prime Minister’s Office, Singapore, under Award NRF2018NCR-NCR009-0001, Award NRF2018NCR-NCR005-0001, Award NRF2018NCR-NSOE003-0001, Award NRFI06-2020-0022, and Award AISG2-RP-2020-019; in part by the Joint Nanyang Technological University (NTU)-WeBank Research Centre on Fintech under Award NWJ-2020-008; in part by Nanyang Assistant Professorship (NAP); and in part by the Research, Innovation and Enterprise (RIE) 2020 Advanced Manufacturing and Engineering Programmatic Fund, Singapore, under Award A20G8b0102. 2024-07-18T00:31:59Z 2024-07-18T00:31:59Z 2021 Journal Article Guo, S., Zhang, T., Yu, H., Xie, X., Ma, L., Xiang, T. & Liu, Y. (2021). Byzantine-resilient decentralized stochastic gradient descent. IEEE Transactions On Circuits and Systems for Video Technology, 32(6), 4096-4106. https://dx.doi.org/10.1109/TCSVT.2021.3116976 1051-8215 https://hdl.handle.net/10356/179057 10.1109/TCSVT.2021.3116976 6 32 4096 4106 en AISG2-RP-2020-019 A20G8b0102 RS02/19 2018-T1-002-069 NRF2018NCR-NCR009-0001 NRF2018NCR-NCR005-0001 NRF2018NCR-NSOE003-0001 NRFI06-2020-0022 NWJ-2020-008 IEEE Transactions on Circuits and Systems for Video Technology © 2021 IEEE. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at http://doi.org/10.1109/TCSVT.2021.3116976. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Artificial intelligence Federated learning |
spellingShingle |
Computer and Information Science Artificial intelligence Federated learning Guo, Shangwei Zhang, Tianwei Yu, Han Xie, Xiaofei Ma, Lei Xiang, Tao Liu, Yang Byzantine-resilient decentralized stochastic gradient descent |
description |
Decentralized learning has gained great popularity to improve learning efficiency and preserve data privacy. Each computing node makes equal contribution to collaboratively learn a Deep Learning model. The elimination of centralized Parameter Servers (PS) can effectively address many issues such as privacy, performance bottleneck and single-point-failure. However, how to achieve Byzantine Fault Tolerance in decentralized learning systems is rarely explored, although this problem has been extensively studied in centralized systems. In this paper, we present an in-depth study towards the Byzantine resilience of decentralized learning systems with two contributions. First, from the adversarial perspective, we theoretically illustrate that Byzantine attacks are more dangerous and feasible in decentralized learning systems: even one malicious participant can arbitrarily alter the models of other participants by sending carefully crafted updates to its neighbors. Second, from the defense perspective, we propose Ubar, a novel algorithm to enhance decentralized learning with Byzantine Fault Tolerance. Specifically, Ubar provides a U niform B yzantine-resilient A ggregation R ule for benign nodes to select the useful parameter updates and filter out the malicious ones in each training iteration. It guarantees that each benign node in a decentralized system can train a correct model under very strong Byzantine attacks with an arbitrary number of faulty nodes. We conduct extensive experiments on standard image classification tasks and the results indicate that Ubar can effectively defeat both simple and sophisticated Byzantine attacks with higher performance efficiency than existing solutions. |
author2 |
College of Computing and Data Science |
author_facet |
College of Computing and Data Science Guo, Shangwei Zhang, Tianwei Yu, Han Xie, Xiaofei Ma, Lei Xiang, Tao Liu, Yang |
format |
Article |
author |
Guo, Shangwei Zhang, Tianwei Yu, Han Xie, Xiaofei Ma, Lei Xiang, Tao Liu, Yang |
author_sort |
Guo, Shangwei |
title |
Byzantine-resilient decentralized stochastic gradient descent |
title_short |
Byzantine-resilient decentralized stochastic gradient descent |
title_full |
Byzantine-resilient decentralized stochastic gradient descent |
title_fullStr |
Byzantine-resilient decentralized stochastic gradient descent |
title_full_unstemmed |
Byzantine-resilient decentralized stochastic gradient descent |
title_sort |
byzantine-resilient decentralized stochastic gradient descent |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/179057 |
_version_ |
1814047165508485120 |