GTG-shapley: efficient and accurate participant contribution evaluation in federated learning

Federated Learning (FL) bridges the gap between collaborative machine learning and preserving data privacy. To sustain the long-term operation of an FL ecosystem, it is important to attract high-quality data owners with appropriate incentive schemes. As an important building block of such incentive...

Full description

Saved in:
Bibliographic Details
Main Authors: Liu, Zelei, Chen, Yuanyuan, Yu, Han, Liu, Yang, Cui, Lizhen
Other Authors: College of Computing and Data Science
Format: Article
Language:English
Published: 2024
Subjects:
Online Access:https://hdl.handle.net/10356/179060
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-179060
record_format dspace
spelling sg-ntu-dr.10356-1790602024-07-18T01:09:40Z GTG-shapley: efficient and accurate participant contribution evaluation in federated learning Liu, Zelei Chen, Yuanyuan Yu, Han Liu, Yang Cui, Lizhen College of Computing and Data Science School of Computer Science and Engineering Computer and Information Science Artificial intelligence Federated learning Federated Learning (FL) bridges the gap between collaborative machine learning and preserving data privacy. To sustain the long-term operation of an FL ecosystem, it is important to attract high-quality data owners with appropriate incentive schemes. As an important building block of such incentive schemes, it is essential to fairly evaluate participants’ contribution to the performance of the final FL model without exposing their private data. Shapley Value (SV)–based techniques have been widely adopted to provide a fair evaluation of FL participant contributions. However, existing approaches incur significant computation costs, making them difficult to apply in practice. In this article, we propose the Guided Truncation Gradient Shapley (GTG-Shapley) approach to address this challenge. It reconstructs FL models from gradient updates for SV calculation instead of repeatedly training with different combinations of FL participants. In addition, we design a guided Monte Carlo sampling approach combined with within-round and between-round truncation to further reduce the number of model reconstructions and evaluations required. This is accomplished through extensive experiments under diverse realistic data distribution settings. The results demonstrate that GTG-Shapley can closely approximate actual Shapley values while significantly increasing computational efficiency compared with the state-of-the-art, especially under non-i.i.d. settings. Agency for Science, Technology and Research (A*STAR) AI Singapore Nanyang Technological University National Research Foundation (NRF) Submitted/Accepted version This research is supported by the National Research Foundation, Singapore, under its AI Singapore Programme (AISG Award No. AISG2-RP-2020-019); the Joint NTU-WeBank Research Centre on Fintech (Award No. NWJ-2020-008), Nanyang Technological University, Singapore; the Nanyang Assistant Professorship (NAP); the RIE 2020 Advanced Manufacturing and Engineering (AME) Programmatic Fund (No. A20G8b0102), Singapore; and the Joint SDU-NTU Research Centre on Artificial Intelligence (C-FAIR), Shandong University, China (NSC-2019-011); NSFC No. 91846205; SDNSFC No. ZR2019LZH008; Shandong Provincial Key Research and Development Program (Major Scientific and Technological Innovation Project, No. 2021CXGC010108). 2024-07-18T01:09:40Z 2024-07-18T01:09:40Z 2022 Journal Article Liu, Z., Chen, Y., Yu, H., Liu, Y. & Cui, L. (2022). GTG-shapley: efficient and accurate participant contribution evaluation in federated learning. ACM Transactions On Intelligent Systems and Technology, 13(4), 60-. https://dx.doi.org/10.1145/3501811 2157-6904 https://hdl.handle.net/10356/179060 10.1145/3501811 4 13 60 en AISG2-RP-2020-019 A20G8b0102) NWJ-2020-008 NSC-2019-011 ACM Transactions on Intelligent Systems and Technology © 2022 Association for Computing Machinery. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at http://doi.org/10.1145/3501811. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
Artificial intelligence
Federated learning
spellingShingle Computer and Information Science
Artificial intelligence
Federated learning
Liu, Zelei
Chen, Yuanyuan
Yu, Han
Liu, Yang
Cui, Lizhen
GTG-shapley: efficient and accurate participant contribution evaluation in federated learning
description Federated Learning (FL) bridges the gap between collaborative machine learning and preserving data privacy. To sustain the long-term operation of an FL ecosystem, it is important to attract high-quality data owners with appropriate incentive schemes. As an important building block of such incentive schemes, it is essential to fairly evaluate participants’ contribution to the performance of the final FL model without exposing their private data. Shapley Value (SV)–based techniques have been widely adopted to provide a fair evaluation of FL participant contributions. However, existing approaches incur significant computation costs, making them difficult to apply in practice. In this article, we propose the Guided Truncation Gradient Shapley (GTG-Shapley) approach to address this challenge. It reconstructs FL models from gradient updates for SV calculation instead of repeatedly training with different combinations of FL participants. In addition, we design a guided Monte Carlo sampling approach combined with within-round and between-round truncation to further reduce the number of model reconstructions and evaluations required. This is accomplished through extensive experiments under diverse realistic data distribution settings. The results demonstrate that GTG-Shapley can closely approximate actual Shapley values while significantly increasing computational efficiency compared with the state-of-the-art, especially under non-i.i.d. settings.
author2 College of Computing and Data Science
author_facet College of Computing and Data Science
Liu, Zelei
Chen, Yuanyuan
Yu, Han
Liu, Yang
Cui, Lizhen
format Article
author Liu, Zelei
Chen, Yuanyuan
Yu, Han
Liu, Yang
Cui, Lizhen
author_sort Liu, Zelei
title GTG-shapley: efficient and accurate participant contribution evaluation in federated learning
title_short GTG-shapley: efficient and accurate participant contribution evaluation in federated learning
title_full GTG-shapley: efficient and accurate participant contribution evaluation in federated learning
title_fullStr GTG-shapley: efficient and accurate participant contribution evaluation in federated learning
title_full_unstemmed GTG-shapley: efficient and accurate participant contribution evaluation in federated learning
title_sort gtg-shapley: efficient and accurate participant contribution evaluation in federated learning
publishDate 2024
url https://hdl.handle.net/10356/179060
_version_ 1814047416514510848