Semi-supervised federated heterogeneous transfer learning

Federated learning (FL) is a privacy-preserving paradigm that collaboratively train machine learning models with distributed data stored in different silos without exposing sensitive information. Different from most existing FL approaches requiring data from different parties share either the same f...

Full description

Saved in:
Bibliographic Details
Main Authors: Feng, Siwei, Li, Boyang, Yu, Han, Liu, Yang, Yang, Qiang
Other Authors: School of Computer Science and Engineering
Format: Article
Language:English
Published: 2022
Subjects:
Online Access:https://hdl.handle.net/10356/163377
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-163377
record_format dspace
spelling sg-ntu-dr.10356-1633772022-12-05T05:09:25Z Semi-supervised federated heterogeneous transfer learning Feng, Siwei Li, Boyang Yu, Han Liu, Yang Yang, Qiang School of Computer Science and Engineering Engineering::Computer science and engineering Federated Transfer Learning Data Privacy Preservation Federated learning (FL) is a privacy-preserving paradigm that collaboratively train machine learning models with distributed data stored in different silos without exposing sensitive information. Different from most existing FL approaches requiring data from different parties share either the same feature space or sample ID space, federated transfer learning (FTL), which is a recently proposed FL concept, is designed for situations where data from different parties differ not only in samples but also in feature space. However, like most traditional FL approaches, FTL methods also suffer from issues caused by insufficiency of overlapping data. In this paper, we propose a novel FTL framework referred to as Semi-Supervised Federated Heterogeneous Transfer Learning (SFHTL) to leverage on the unlabeled non-overlapping samples to reduce model overfitting as a result of insufficient overlapping training samples in FL scenarios. Unlike existing FTL approaches, SFHTL makes use of non-overlapping samples from all parties to expand the training set for each party to improve local model performance. Through extensive experimental evaluation based on real-world datasets, we demonstrate significant advantages of SFHTL over state-of-the-art approaches. Nanyang Technological University National Research Foundation (NRF) This research is supported, in part, by the National Natural Science Foundation of China under Grant NSFC 62106167; Project Funded by the Priority Academic Program Development of Jiangsu Higher Education Institutions; the National Research Foundation, Singapore under its AI Singapore Programme (AISG Award No: AISG2-RP-2020-019); the RIE 2020 Advanced Manufacturing and Engineering (AME) Programmatic Fund (No. A20G8b0102), Singapore; Nanyang Assistant Professorship (NAP); the Joint NTUWeBank Research Centre on Fintech (Award No: NWJ-2020-008), Nanyang Technological University, Singapore; and Joint SDU-NTU Centre for Artificial Intelligence Research (C-FAIR) (NSC-2019- 011). Qiang Yang is supported in part by China National Key Research and Development Program of China under Grant No. 2018AAA0101100. 2022-12-05T05:09:24Z 2022-12-05T05:09:24Z 2022 Journal Article Feng, S., Li, B., Yu, H., Liu, Y. & Yang, Q. (2022). Semi-supervised federated heterogeneous transfer learning. Knowledge-Based Systems, 252, 109384-. https://dx.doi.org/10.1016/j.knosys.2022.109384 0950-7051 https://hdl.handle.net/10356/163377 10.1016/j.knosys.2022.109384 2-s2.0-85134631495 252 109384 en AISG2-RP-2020-019 A20G8b0102 NWJ-2020-008 NSC-2019-011 Knowledge-Based Systems © 2022 Elsevier B.V. All rights reserved.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering
Federated Transfer Learning
Data Privacy Preservation
spellingShingle Engineering::Computer science and engineering
Federated Transfer Learning
Data Privacy Preservation
Feng, Siwei
Li, Boyang
Yu, Han
Liu, Yang
Yang, Qiang
Semi-supervised federated heterogeneous transfer learning
description Federated learning (FL) is a privacy-preserving paradigm that collaboratively train machine learning models with distributed data stored in different silos without exposing sensitive information. Different from most existing FL approaches requiring data from different parties share either the same feature space or sample ID space, federated transfer learning (FTL), which is a recently proposed FL concept, is designed for situations where data from different parties differ not only in samples but also in feature space. However, like most traditional FL approaches, FTL methods also suffer from issues caused by insufficiency of overlapping data. In this paper, we propose a novel FTL framework referred to as Semi-Supervised Federated Heterogeneous Transfer Learning (SFHTL) to leverage on the unlabeled non-overlapping samples to reduce model overfitting as a result of insufficient overlapping training samples in FL scenarios. Unlike existing FTL approaches, SFHTL makes use of non-overlapping samples from all parties to expand the training set for each party to improve local model performance. Through extensive experimental evaluation based on real-world datasets, we demonstrate significant advantages of SFHTL over state-of-the-art approaches.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Feng, Siwei
Li, Boyang
Yu, Han
Liu, Yang
Yang, Qiang
format Article
author Feng, Siwei
Li, Boyang
Yu, Han
Liu, Yang
Yang, Qiang
author_sort Feng, Siwei
title Semi-supervised federated heterogeneous transfer learning
title_short Semi-supervised federated heterogeneous transfer learning
title_full Semi-supervised federated heterogeneous transfer learning
title_fullStr Semi-supervised federated heterogeneous transfer learning
title_full_unstemmed Semi-supervised federated heterogeneous transfer learning
title_sort semi-supervised federated heterogeneous transfer learning
publishDate 2022
url https://hdl.handle.net/10356/163377
_version_ 1751548556157648896