Decentralized federated learning

Conventional implementations of federated learning require a centralized entity to conduct and coordinate the training with a star communication architecture. However, this technique is prone to a single point of failure, e.g., when the central node is malicious. In this study, we explore decen...

Full description

Saved in:
Bibliographic Details
Main Author: Hitesh, Agarwal
Other Authors: Dusit Niyato
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2022
Subjects:
Online Access:https://hdl.handle.net/10356/156589
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-156589
record_format dspace
spelling sg-ntu-dr.10356-1565892022-04-20T08:39:44Z Decentralized federated learning Hitesh, Agarwal Dusit Niyato School of Computer Science and Engineering DNIYATO@ntu.edu.sg Engineering::Computer science and engineering Conventional implementations of federated learning require a centralized entity to conduct and coordinate the training with a star communication architecture. However, this technique is prone to a single point of failure, e.g., when the central node is malicious. In this study, we explore decentralized federated learning frameworks where clients communicate with each other following a peer-to-peer mechanism rather than server-client. We study how communication topology and model partitioning affects the throughput and convergence metrics in decentralized federated learning. To make our study as practically applicable as possible, we include network link latencies in our performance metrics for a fair evaluation. Through our study, we conclude that the ring communication mechanism has the highest throughput with the best convergence performance metrics. In big networks, ring is almost 8 times as fast as centralized communications. Bachelor of Science in Data Science and Artificial Intelligence 2022-04-20T08:39:44Z 2022-04-20T08:39:44Z 2022 Final Year Project (FYP) Hitesh, A. (2022). Decentralized federated learning. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/156589 https://hdl.handle.net/10356/156589 en SCSE21-0200 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering
spellingShingle Engineering::Computer science and engineering
Hitesh, Agarwal
Decentralized federated learning
description Conventional implementations of federated learning require a centralized entity to conduct and coordinate the training with a star communication architecture. However, this technique is prone to a single point of failure, e.g., when the central node is malicious. In this study, we explore decentralized federated learning frameworks where clients communicate with each other following a peer-to-peer mechanism rather than server-client. We study how communication topology and model partitioning affects the throughput and convergence metrics in decentralized federated learning. To make our study as practically applicable as possible, we include network link latencies in our performance metrics for a fair evaluation. Through our study, we conclude that the ring communication mechanism has the highest throughput with the best convergence performance metrics. In big networks, ring is almost 8 times as fast as centralized communications.
author2 Dusit Niyato
author_facet Dusit Niyato
Hitesh, Agarwal
format Final Year Project
author Hitesh, Agarwal
author_sort Hitesh, Agarwal
title Decentralized federated learning
title_short Decentralized federated learning
title_full Decentralized federated learning
title_fullStr Decentralized federated learning
title_full_unstemmed Decentralized federated learning
title_sort decentralized federated learning
publisher Nanyang Technological University
publishDate 2022
url https://hdl.handle.net/10356/156589
_version_ 1731235747873161216