CaBaFL: Asynchronous federated learning via hierarchical cache and feature balance

Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massi...

Full description

Saved in:
Bibliographic Details
Main Authors: XIA, Zeke, HU, Ming, YAN, Dengke, XIE, Xiaofei, LI, Tianlin, LI, Anran, ZHOU, Junlong, CHEN, Mingsong
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2024
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/9563
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-10563
record_format dspace
spelling sg-smu-ink.sis_research-105632024-11-15T06:54:03Z CaBaFL: Asynchronous federated learning via hierarchical cache and feature balance XIA, Zeke HU, Ming YAN, Dengke XIE, Xiaofei LI, Tianlin LI, Anran ZHOU, Junlong CHEN, Mingsong Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massive AIoT devices, respectively. To address the above challenges, we present a novel asynchronous FL approach named CaBaFL, which includes a hierarchical cache-based aggregation mechanism and a feature balance-guided device selection strategy. CaBaFL maintains multiple intermediate models simultaneously for local training. The hierarchical cache-based aggregation mechanism enables each intermediate model to be trained on multiple devices to align the training time and mitigate the straggler issue. In specific, each intermediate model is stored in a low-level cache for local training and when it is trained by sufficient local devices, it will be stored in a high-level cache for aggregation. To address the problem of imbalanced data, the feature balance-guided device selection strategy in CaBaFL adopts the activation distribution as a metric, which enables each intermediate model to be trained across devices with totally balanced data distributions before aggregation. Experimental results show that compared to the state-of-the-art FL methods, CaBaFL achieves up to 9.26X training acceleration and 19.71% accuracy improvements. 2024-11-01T07:00:00Z text https://ink.library.smu.edu.sg/sis_research/9563 info:doi/10.1109/TCAD.2024.3446881 Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Artificial Intelligence of Things AIoT) asynchronous federated learning data/device heterogeneity feature balance Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Artificial Intelligence of Things AIoT)
asynchronous federated learning
data/device heterogeneity
feature balance
Software Engineering
spellingShingle Artificial Intelligence of Things AIoT)
asynchronous federated learning
data/device heterogeneity
feature balance
Software Engineering
XIA, Zeke
HU, Ming
YAN, Dengke
XIE, Xiaofei
LI, Tianlin
LI, Anran
ZHOU, Junlong
CHEN, Mingsong
CaBaFL: Asynchronous federated learning via hierarchical cache and feature balance
description Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massive AIoT devices, respectively. To address the above challenges, we present a novel asynchronous FL approach named CaBaFL, which includes a hierarchical cache-based aggregation mechanism and a feature balance-guided device selection strategy. CaBaFL maintains multiple intermediate models simultaneously for local training. The hierarchical cache-based aggregation mechanism enables each intermediate model to be trained on multiple devices to align the training time and mitigate the straggler issue. In specific, each intermediate model is stored in a low-level cache for local training and when it is trained by sufficient local devices, it will be stored in a high-level cache for aggregation. To address the problem of imbalanced data, the feature balance-guided device selection strategy in CaBaFL adopts the activation distribution as a metric, which enables each intermediate model to be trained across devices with totally balanced data distributions before aggregation. Experimental results show that compared to the state-of-the-art FL methods, CaBaFL achieves up to 9.26X training acceleration and 19.71% accuracy improvements.
format text
author XIA, Zeke
HU, Ming
YAN, Dengke
XIE, Xiaofei
LI, Tianlin
LI, Anran
ZHOU, Junlong
CHEN, Mingsong
author_facet XIA, Zeke
HU, Ming
YAN, Dengke
XIE, Xiaofei
LI, Tianlin
LI, Anran
ZHOU, Junlong
CHEN, Mingsong
author_sort XIA, Zeke
title CaBaFL: Asynchronous federated learning via hierarchical cache and feature balance
title_short CaBaFL: Asynchronous federated learning via hierarchical cache and feature balance
title_full CaBaFL: Asynchronous federated learning via hierarchical cache and feature balance
title_fullStr CaBaFL: Asynchronous federated learning via hierarchical cache and feature balance
title_full_unstemmed CaBaFL: Asynchronous federated learning via hierarchical cache and feature balance
title_sort cabafl: asynchronous federated learning via hierarchical cache and feature balance
publisher Institutional Knowledge at Singapore Management University
publishDate 2024
url https://ink.library.smu.edu.sg/sis_research/9563
_version_ 1816859133780951040