Cargando…

Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond

Machine learning (ML) in healthcare data analytics is attracting much attention because of the unprecedented power of ML to extract knowledge that improves the decision-making process. At the same time, laws and ethics codes drafted by countries to govern healthcare data are becoming stringent. Alth...

Descripción completa

Detalles Bibliográficos
Autores principales: Joshi, Praveen, Thapa, Chandra, Camtepe, Seyit, Hasanuzzaman, Mohammed, Scully, Ted, Afli, Haithem
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9326525/
https://www.ncbi.nlm.nih.gov/pubmed/35893586
http://dx.doi.org/10.3390/mps5040060
_version_ 1784757305959514112
author Joshi, Praveen
Thapa, Chandra
Camtepe, Seyit
Hasanuzzaman, Mohammed
Scully, Ted
Afli, Haithem
author_facet Joshi, Praveen
Thapa, Chandra
Camtepe, Seyit
Hasanuzzaman, Mohammed
Scully, Ted
Afli, Haithem
author_sort Joshi, Praveen
collection PubMed
description Machine learning (ML) in healthcare data analytics is attracting much attention because of the unprecedented power of ML to extract knowledge that improves the decision-making process. At the same time, laws and ethics codes drafted by countries to govern healthcare data are becoming stringent. Although healthcare practitioners are struggling with an enforced governance framework, we see the emergence of distributed learning-based frameworks disrupting traditional-ML-model development. Splitfed learning (SFL) is one of the recent developments in distributed machine learning that empowers healthcare practitioners to preserve the privacy of input data and enables them to train ML models. However, SFL has some extra communication and computation overheads at the client side due to the requirement of client-side model synchronization. For a resource-constrained client side (hospitals with limited computational powers), removing such conditions is required to gain efficiency in the learning. In this regard, this paper studies SFL without client-side model synchronization. The resulting architecture is known as multi-head split learning (MHSL). At the same time, it is important to investigate information leakage, which indicates how much information is gained by the server related to the raw data directly out of the smashed data—the output of the client-side model portion—passed to it by the client. Our empirical studies examine the Resnet-18 and Conv1-D architecture model on the ECG and HAM-10000 datasets under IID data distribution. The results find that SFL provides 1.81% and 2.36% better accuracy than MHSL on the ECG and HAM-10000 datasets, respectively (for cut-layer value set to 1). Analysis of experimentation with various client-side model portions demonstrates that it has an impact on the overall performance. With an increase in layers in the client-side model portion, SFL performance improves while MHSL performance degrades. Experiment results also demonstrate that information leakage provided by mutual information score values in SFL is more than MHSL for ECG and HAM-10000 datasets by [Formula: see text] and [Formula: see text] , respectively.
format Online
Article
Text
id pubmed-9326525
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-93265252022-07-28 Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond Joshi, Praveen Thapa, Chandra Camtepe, Seyit Hasanuzzaman, Mohammed Scully, Ted Afli, Haithem Methods Protoc Article Machine learning (ML) in healthcare data analytics is attracting much attention because of the unprecedented power of ML to extract knowledge that improves the decision-making process. At the same time, laws and ethics codes drafted by countries to govern healthcare data are becoming stringent. Although healthcare practitioners are struggling with an enforced governance framework, we see the emergence of distributed learning-based frameworks disrupting traditional-ML-model development. Splitfed learning (SFL) is one of the recent developments in distributed machine learning that empowers healthcare practitioners to preserve the privacy of input data and enables them to train ML models. However, SFL has some extra communication and computation overheads at the client side due to the requirement of client-side model synchronization. For a resource-constrained client side (hospitals with limited computational powers), removing such conditions is required to gain efficiency in the learning. In this regard, this paper studies SFL without client-side model synchronization. The resulting architecture is known as multi-head split learning (MHSL). At the same time, it is important to investigate information leakage, which indicates how much information is gained by the server related to the raw data directly out of the smashed data—the output of the client-side model portion—passed to it by the client. Our empirical studies examine the Resnet-18 and Conv1-D architecture model on the ECG and HAM-10000 datasets under IID data distribution. The results find that SFL provides 1.81% and 2.36% better accuracy than MHSL on the ECG and HAM-10000 datasets, respectively (for cut-layer value set to 1). Analysis of experimentation with various client-side model portions demonstrates that it has an impact on the overall performance. With an increase in layers in the client-side model portion, SFL performance improves while MHSL performance degrades. Experiment results also demonstrate that information leakage provided by mutual information score values in SFL is more than MHSL for ECG and HAM-10000 datasets by [Formula: see text] and [Formula: see text] , respectively. MDPI 2022-07-13 /pmc/articles/PMC9326525/ /pubmed/35893586 http://dx.doi.org/10.3390/mps5040060 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Joshi, Praveen
Thapa, Chandra
Camtepe, Seyit
Hasanuzzaman, Mohammed
Scully, Ted
Afli, Haithem
Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond
title Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond
title_full Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond
title_fullStr Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond
title_full_unstemmed Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond
title_short Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond
title_sort performance and information leakage in splitfed learning and multi-head split learning in healthcare data and beyond
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9326525/
https://www.ncbi.nlm.nih.gov/pubmed/35893586
http://dx.doi.org/10.3390/mps5040060
work_keys_str_mv AT joshipraveen performanceandinformationleakageinsplitfedlearningandmultiheadsplitlearninginhealthcaredataandbeyond
AT thapachandra performanceandinformationleakageinsplitfedlearningandmultiheadsplitlearninginhealthcaredataandbeyond
AT camtepeseyit performanceandinformationleakageinsplitfedlearningandmultiheadsplitlearninginhealthcaredataandbeyond
AT hasanuzzamanmohammed performanceandinformationleakageinsplitfedlearningandmultiheadsplitlearninginhealthcaredataandbeyond
AT scullyted performanceandinformationleakageinsplitfedlearningandmultiheadsplitlearninginhealthcaredataandbeyond
AT aflihaithem performanceandinformationleakageinsplitfedlearningandmultiheadsplitlearninginhealthcaredataandbeyond