GTG-shapley: efficient and accurate participant contribution evaluation in federated learning
Federated Learning (FL) bridges the gap between collaborative machine learning and preserving data privacy. To sustain the long-term operation of an FL ecosystem, it is important to attract high-quality data owners with appropriate incentive schemes. As an important building block of such incentive...
Saved in:
Main Authors: | , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/179060 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-179060 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1790602024-07-18T01:09:40Z GTG-shapley: efficient and accurate participant contribution evaluation in federated learning Liu, Zelei Chen, Yuanyuan Yu, Han Liu, Yang Cui, Lizhen College of Computing and Data Science School of Computer Science and Engineering Computer and Information Science Artificial intelligence Federated learning Federated Learning (FL) bridges the gap between collaborative machine learning and preserving data privacy. To sustain the long-term operation of an FL ecosystem, it is important to attract high-quality data owners with appropriate incentive schemes. As an important building block of such incentive schemes, it is essential to fairly evaluate participants’ contribution to the performance of the final FL model without exposing their private data. Shapley Value (SV)–based techniques have been widely adopted to provide a fair evaluation of FL participant contributions. However, existing approaches incur significant computation costs, making them difficult to apply in practice. In this article, we propose the Guided Truncation Gradient Shapley (GTG-Shapley) approach to address this challenge. It reconstructs FL models from gradient updates for SV calculation instead of repeatedly training with different combinations of FL participants. In addition, we design a guided Monte Carlo sampling approach combined with within-round and between-round truncation to further reduce the number of model reconstructions and evaluations required. This is accomplished through extensive experiments under diverse realistic data distribution settings. The results demonstrate that GTG-Shapley can closely approximate actual Shapley values while significantly increasing computational efficiency compared with the state-of-the-art, especially under non-i.i.d. settings. Agency for Science, Technology and Research (A*STAR) AI Singapore Nanyang Technological University National Research Foundation (NRF) Submitted/Accepted version This research is supported by the National Research Foundation, Singapore, under its AI Singapore Programme (AISG Award No. AISG2-RP-2020-019); the Joint NTU-WeBank Research Centre on Fintech (Award No. NWJ-2020-008), Nanyang Technological University, Singapore; the Nanyang Assistant Professorship (NAP); the RIE 2020 Advanced Manufacturing and Engineering (AME) Programmatic Fund (No. A20G8b0102), Singapore; and the Joint SDU-NTU Research Centre on Artificial Intelligence (C-FAIR), Shandong University, China (NSC-2019-011); NSFC No. 91846205; SDNSFC No. ZR2019LZH008; Shandong Provincial Key Research and Development Program (Major Scientific and Technological Innovation Project, No. 2021CXGC010108). 2024-07-18T01:09:40Z 2024-07-18T01:09:40Z 2022 Journal Article Liu, Z., Chen, Y., Yu, H., Liu, Y. & Cui, L. (2022). GTG-shapley: efficient and accurate participant contribution evaluation in federated learning. ACM Transactions On Intelligent Systems and Technology, 13(4), 60-. https://dx.doi.org/10.1145/3501811 2157-6904 https://hdl.handle.net/10356/179060 10.1145/3501811 4 13 60 en AISG2-RP-2020-019 A20G8b0102) NWJ-2020-008 NSC-2019-011 ACM Transactions on Intelligent Systems and Technology © 2022 Association for Computing Machinery. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at http://doi.org/10.1145/3501811. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Artificial intelligence Federated learning |
spellingShingle |
Computer and Information Science Artificial intelligence Federated learning Liu, Zelei Chen, Yuanyuan Yu, Han Liu, Yang Cui, Lizhen GTG-shapley: efficient and accurate participant contribution evaluation in federated learning |
description |
Federated Learning (FL) bridges the gap between collaborative machine learning and preserving data privacy. To sustain the long-term operation of an FL ecosystem, it is important to attract high-quality data owners with appropriate incentive schemes. As an important building block of such incentive schemes, it is essential to fairly evaluate participants’ contribution to the performance of the final FL model without exposing their private data. Shapley Value (SV)–based techniques have been widely adopted to provide a fair evaluation of FL participant contributions. However, existing approaches incur significant computation costs, making them difficult to apply in practice. In this article, we propose the Guided Truncation Gradient Shapley (GTG-Shapley) approach to address this challenge. It reconstructs FL models from gradient updates for SV calculation instead of repeatedly training with different combinations of FL participants. In addition, we design a guided Monte Carlo sampling approach combined with within-round and between-round truncation to further reduce the number of model reconstructions and evaluations required. This is accomplished through extensive experiments under diverse realistic data distribution settings. The results demonstrate that GTG-Shapley can closely approximate actual Shapley values while significantly increasing computational efficiency compared with the state-of-the-art, especially under non-i.i.d. settings. |
author2 |
College of Computing and Data Science |
author_facet |
College of Computing and Data Science Liu, Zelei Chen, Yuanyuan Yu, Han Liu, Yang Cui, Lizhen |
format |
Article |
author |
Liu, Zelei Chen, Yuanyuan Yu, Han Liu, Yang Cui, Lizhen |
author_sort |
Liu, Zelei |
title |
GTG-shapley: efficient and accurate participant contribution evaluation in federated learning |
title_short |
GTG-shapley: efficient and accurate participant contribution evaluation in federated learning |
title_full |
GTG-shapley: efficient and accurate participant contribution evaluation in federated learning |
title_fullStr |
GTG-shapley: efficient and accurate participant contribution evaluation in federated learning |
title_full_unstemmed |
GTG-shapley: efficient and accurate participant contribution evaluation in federated learning |
title_sort |
gtg-shapley: efficient and accurate participant contribution evaluation in federated learning |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/179060 |
_version_ |
1814047416514510848 |