Byzantine-resilient decentralized stochastic gradient descent

Decentralized learning has gained great popularity to improve learning efficiency and preserve data privacy. Each computing node makes equal contribution to collaboratively learn a Deep Learning model. The elimination of centralized Parameter Servers (PS) can effectively address many issues such as...

Full description

Saved in:
Bibliographic Details
Main Authors: Guo, Shangwei, Zhang, Tianwei, Yu, Han, Xie, Xiaofei, Ma, Lei, Xiang, Tao, Liu, Yang
Other Authors: College of Computing and Data Science
Format: Article
Language:English
Published: 2024
Subjects:
Online Access:https://hdl.handle.net/10356/179057
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-179057
record_format dspace
spelling sg-ntu-dr.10356-1790572024-07-18T00:32:00Z Byzantine-resilient decentralized stochastic gradient descent Guo, Shangwei Zhang, Tianwei Yu, Han Xie, Xiaofei Ma, Lei Xiang, Tao Liu, Yang College of Computing and Data Science School of Computer Science and Engineering Computer and Information Science Artificial intelligence Federated learning Decentralized learning has gained great popularity to improve learning efficiency and preserve data privacy. Each computing node makes equal contribution to collaboratively learn a Deep Learning model. The elimination of centralized Parameter Servers (PS) can effectively address many issues such as privacy, performance bottleneck and single-point-failure. However, how to achieve Byzantine Fault Tolerance in decentralized learning systems is rarely explored, although this problem has been extensively studied in centralized systems. In this paper, we present an in-depth study towards the Byzantine resilience of decentralized learning systems with two contributions. First, from the adversarial perspective, we theoretically illustrate that Byzantine attacks are more dangerous and feasible in decentralized learning systems: even one malicious participant can arbitrarily alter the models of other participants by sending carefully crafted updates to its neighbors. Second, from the defense perspective, we propose Ubar, a novel algorithm to enhance decentralized learning with Byzantine Fault Tolerance. Specifically, Ubar provides a U niform B yzantine-resilient A ggregation R ule for benign nodes to select the useful parameter updates and filter out the malicious ones in each training iteration. It guarantees that each benign node in a decentralized system can train a correct model under very strong Byzantine attacks with an arbitrary number of faulty nodes. We conduct extensive experiments on standard image classification tasks and the results indicate that Ubar can effectively defeat both simple and sophisticated Byzantine attacks with higher performance efficiency than existing solutions. Agency for Science, Technology and Research (A*STAR) AI Singapore Ministry of Education (MOE) Nanyang Technological University National Research Foundation (NRF) Submitted/Accepted version This work was supported in part by the National Natural Science Foundation of China under Grant 62102052; in part by Singapore Ministry of Education Academic Research Fund Tier 1 under Award RS02/19 and Award 2018-T1-002-069; in part by the National Research Foundation, Prime Minister’s Office, Singapore, under Award NRF2018NCR-NCR009-0001, Award NRF2018NCR-NCR005-0001, Award NRF2018NCR-NSOE003-0001, Award NRFI06-2020-0022, and Award AISG2-RP-2020-019; in part by the Joint Nanyang Technological University (NTU)-WeBank Research Centre on Fintech under Award NWJ-2020-008; in part by Nanyang Assistant Professorship (NAP); and in part by the Research, Innovation and Enterprise (RIE) 2020 Advanced Manufacturing and Engineering Programmatic Fund, Singapore, under Award A20G8b0102. 2024-07-18T00:31:59Z 2024-07-18T00:31:59Z 2021 Journal Article Guo, S., Zhang, T., Yu, H., Xie, X., Ma, L., Xiang, T. & Liu, Y. (2021). Byzantine-resilient decentralized stochastic gradient descent. IEEE Transactions On Circuits and Systems for Video Technology, 32(6), 4096-4106. https://dx.doi.org/10.1109/TCSVT.2021.3116976 1051-8215 https://hdl.handle.net/10356/179057 10.1109/TCSVT.2021.3116976 6 32 4096 4106 en AISG2-RP-2020-019 A20G8b0102 RS02/19 2018-T1-002-069 NRF2018NCR-NCR009-0001 NRF2018NCR-NCR005-0001 NRF2018NCR-NSOE003-0001 NRFI06-2020-0022 NWJ-2020-008 IEEE Transactions on Circuits and Systems for Video Technology © 2021 IEEE. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at http://doi.org/10.1109/TCSVT.2021.3116976. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
Artificial intelligence
Federated learning
spellingShingle Computer and Information Science
Artificial intelligence
Federated learning
Guo, Shangwei
Zhang, Tianwei
Yu, Han
Xie, Xiaofei
Ma, Lei
Xiang, Tao
Liu, Yang
Byzantine-resilient decentralized stochastic gradient descent
description Decentralized learning has gained great popularity to improve learning efficiency and preserve data privacy. Each computing node makes equal contribution to collaboratively learn a Deep Learning model. The elimination of centralized Parameter Servers (PS) can effectively address many issues such as privacy, performance bottleneck and single-point-failure. However, how to achieve Byzantine Fault Tolerance in decentralized learning systems is rarely explored, although this problem has been extensively studied in centralized systems. In this paper, we present an in-depth study towards the Byzantine resilience of decentralized learning systems with two contributions. First, from the adversarial perspective, we theoretically illustrate that Byzantine attacks are more dangerous and feasible in decentralized learning systems: even one malicious participant can arbitrarily alter the models of other participants by sending carefully crafted updates to its neighbors. Second, from the defense perspective, we propose Ubar, a novel algorithm to enhance decentralized learning with Byzantine Fault Tolerance. Specifically, Ubar provides a U niform B yzantine-resilient A ggregation R ule for benign nodes to select the useful parameter updates and filter out the malicious ones in each training iteration. It guarantees that each benign node in a decentralized system can train a correct model under very strong Byzantine attacks with an arbitrary number of faulty nodes. We conduct extensive experiments on standard image classification tasks and the results indicate that Ubar can effectively defeat both simple and sophisticated Byzantine attacks with higher performance efficiency than existing solutions.
author2 College of Computing and Data Science
author_facet College of Computing and Data Science
Guo, Shangwei
Zhang, Tianwei
Yu, Han
Xie, Xiaofei
Ma, Lei
Xiang, Tao
Liu, Yang
format Article
author Guo, Shangwei
Zhang, Tianwei
Yu, Han
Xie, Xiaofei
Ma, Lei
Xiang, Tao
Liu, Yang
author_sort Guo, Shangwei
title Byzantine-resilient decentralized stochastic gradient descent
title_short Byzantine-resilient decentralized stochastic gradient descent
title_full Byzantine-resilient decentralized stochastic gradient descent
title_fullStr Byzantine-resilient decentralized stochastic gradient descent
title_full_unstemmed Byzantine-resilient decentralized stochastic gradient descent
title_sort byzantine-resilient decentralized stochastic gradient descent
publishDate 2024
url https://hdl.handle.net/10356/179057
_version_ 1814047165508485120