| Literature DB >> 35893586 |
Praveen Joshi1, Chandra Thapa2, Seyit Camtepe2, Mohammed Hasanuzzaman1, Ted Scully1, Haithem Afli1.
Abstract
Machine learning (ML) in healthcare data analytics is attracting much attention because of the unprecedented power of ML to extract knowledge that improves the decision-making process. At the same time, laws and ethics codes drafted by countries to govern healthcare data are becoming stringent. Although healthcare practitioners are struggling with an enforced governance framework, we see the emergence of distributed learning-based frameworks disrupting traditional-ML-model development. Splitfed learning (SFL) is one of the recent developments in distributed machine learning that empowers healthcare practitioners to preserve the privacy of input data and enables them to train ML models. However, SFL has some extra communication and computation overheads at the client side due to the requirement of client-side model synchronization. For a resource-constrained client side (hospitals with limited computational powers), removing such conditions is required to gain efficiency in the learning. In this regard, this paper studies SFL without client-side model synchronization. The resulting architecture is known as multi-head split learning (MHSL). At the same time, it is important to investigate information leakage, which indicates how much information is gained by the server related to the raw data directly out of the smashed data-the output of the client-side model portion-passed to it by the client. Our empirical studies examine the Resnet-18 and Conv1-D architecture model on the ECG and HAM-10000 datasets under IID data distribution. The results find that SFL provides 1.81% and 2.36% better accuracy than MHSL on the ECG and HAM-10000 datasets, respectively (for cut-layer value set to 1). Analysis of experimentation with various client-side model portions demonstrates that it has an impact on the overall performance. With an increase in layers in the client-side model portion, SFL performance improves while MHSL performance degrades. Experiment results also demonstrate that information leakage provided by mutual information score values in SFL is more than MHSL for ECG and HAM-10000 datasets by 2×10-5 and 4×10-3, respectively.Entities:
Keywords: distributed collaborative machine learning; information leakage in distributed learning; multi-head split learning; parameter transmission-based distributed machine learning; privacy-preserving machine learning; split learning
Year: 2022 PMID: 35893586 PMCID: PMC9326525 DOI: 10.3390/mps5040060
Source DB: PubMed Journal: Methods Protoc ISSN: 2409-9279
Figure 1Multi-head-split-learning architecture.
Datasets used in our experiment setup.
| Dataset | Training Samples | Testing Samples | Image Size | Dataset Type | Number of Labels |
|---|---|---|---|---|---|
| ECG | 13,245 | 13,245 | NA | Time series Dataset | 5 |
| HAM-10000 | 9013 | 1002 |
| Image Dataset | 7 |
| MNIST | 60,000 | 10,000 |
| Image Dataset | 10 |
| KMNIST | 60,000 | 10,000 |
| Image Dataset | 10 |
| CIFAR-10 | 50,000 | 10,000 |
| Image Dataset | 10 |
Model architecture used in the experimental setup.
| Architecture | No. of Parameters | Layers | Kernel Size |
|---|---|---|---|
| Resnet-18 [ | 18 |
| |
| Conv1-D architecture [ | 55,989 | 8 |
|
Communication size and model-training-time equations for SFL and MHSL for one global epoch.
| Method | Comms. Size per Client | Total Comms. Size | Total Model Training Time |
|---|---|---|---|
| SFL |
|
|
|
| MHSL |
|
|
|
Figure 2Computation time (in seconds) for SFL and MHSL.
Training and testing accuracy for centralized architecture.
| Dataset | Model | Testing Accuracy | Training Accuracy |
|---|---|---|---|
| ECG | Conv1-D architecture | 83.56 | 81.72 |
| HAM-10000 | Resnet-18 | 74.67 | 80.26 |
| MNIST | Resnet-18 | 99.15 | 99.31 |
| KMNIST | Resnet-18 | 95.74 | 99.31 |
| CIFAR-10 | Resnet-18 | 78.02 | 97.52 |
Training and testing accuracy for centralized, SFL and MHSL architectures.
| Dataset | Testing Accuracy | Training Accuracy | |||||
|---|---|---|---|---|---|---|---|
| Centralized | SFL | MHSL | Centralized | SFL | MHSL | ||
| ECG | 83.56 | 81.37 | 79.56 | −1.81 | 81.72 | 80.98 | 79.43 |
| HAM-10000 | 74.67 | 73.40 | 71.04 | −2.36 | 80.26 | 78.80 | 77.77 |
| MNIST | 99.15 | 98.50 | 98.99 | 0.49 | 99.31 | 98.99 | 99.11 |
| KMNIST | 95.74 | 96.23 | 96.17 | −0.06 | 99.31 | 98.98 | 98.99 |
| CIFAR-10 | 78.02 | 76.25 | 73.75 | −2.5 | 97.52 | 97.10 | 96.95 |
Test accuracy with respect to the model split at different layers.
| Dataset | Architecture ↓ Split at Layer → | L1 | L2 | L3 | L4 | L5 | L6 | L7 | L8 | L9 |
|---|---|---|---|---|---|---|---|---|---|---|
| ECG |
| 81.37 | 73.45 | 76.57 | 83.90 | - | - | - | - | - |
| ECG |
| 79.56 | 83.78 | 86.44 | 80.03 | - | - | - | - | - |
| HAM-10000 |
| 73.40 | 77.25 | 76.59 | 75.46 | 76.45 | 75.43 | 76.35 | 70.89 | 76.69 |
| HAM-10000 |
| 71.04 | 71.63 | 73.26 | 71.17 | 72.18 | 72.53 | 71.63 | 71.73 | 68.47 |
| MNIST |
| 98.50 | 98.94 | 99.04 | 99.06 | 99.11 | 99.31 | 99.28 | 99.20 | 99.21 |
| MNIST |
| 98.99 | 98.97 | 98.92 | 98.43 | 98.50 | 98.37 | 98.19 | 98.15 | 98.18 |
| KMNIST |
| 96.23 | 96.56 | 96.17 | 96.60 | 96.57 | 96.83 | 96.36 | 96.61 | 97.13 |
| KMNIST |
| 96.17 | 96.07 | 95.60 | 95.11 | 93.89 | 92.56 | 92.87 | 92.77 | 91.86 |
| CIFAR-10 |
| 76.25 | 76.10 | 75.73 | 76.76 | 76.72 | 77.60 | 78.06 | 79.04 | 78.82 |
| CIFAR-10 |
| 73.75 | 72.37 | 70.94 | 66.83 | 66.47 | 64.91 | 64.58 | 65.70 | 65.58 |
Figure 3Mutual information score across the epochs for SFL and MHSL for the ECG dataset.
Figure 4Mutual information score across the epochs for SFL and MHSL for three-channel datasets (a) HAM-10000 and (b) CIFAR-10.
Figure 5Mutual information score across the epochs for SFL and MHSL for one-channel datasets (a) MNIST and (b) KMNIST.
Visual comparison of input images against SFL and MHSL at cut-layer three during an evaluation phase.
| Dataset | Input Image | SFL | MHSL |
|---|---|---|---|
| HAM-10000 |
|
|
|
| CIFAR-10 |
|
|
|
| MNIST |
|
|
|
| KMNIST |
|
|
|