| Literature DB >> 32369025 |
Sven Festag1,2, Cord Spreckelsen1,2.
Abstract
BACKGROUND: Collaborative privacy-preserving training methods allow for the integration of locally stored private data sets into machine learning approaches while ensuring confidentiality and nondisclosure.Entities:
Keywords: distributed machine learning; health informatics; neural networks; privacy-preserving protocols
Year: 2020 PMID: 32369025 PMCID: PMC7238077 DOI: 10.2196/14064
Source DB: PubMed Journal: JMIR Form Res ISSN: 2561-326X
Possible protected health information classes of terms.
| Class | Subclasses |
| Name | Patient, clinician, username |
| Profession | — |
| Location | Hospital, organization, street, city, state, country, zip, other |
| Age | — |
| Date | — |
| Contact | Phone, fax, email, URL, Internet Protocol addressa |
| Identification | Social security numbera, medical record number, health plan number, account numbera, license numbera, vehicle identificationa, device identification, biometric identification, identification number |
aClasses that are not present in the published data set [7].
Figure 1Network topology of the recurrent neural network used for deidentification [3].
Figure 2Nonprotected training with shared data.
Figure 3Semiprotective round robin training with local data.
Figure 4Distributed selective stochastic gradient descent with local data.
Figure 5Local training procedure [4].
Summary of the experiments.
| Name | Learning strategy |
| A | Centralized training using stochastic gradient descent (learning rate: 0 |
| B | Collaborative training of 5 workers using the round robin method (learning rate: 0 |
| C_0 | Collaborative privacy preserving training of 5 workers with DSSGDa (θd = 0.1, θu = 0.5, γ = 10, τ = 0.0001; learning rate: 0.9) |
| C_0 | Collaborative privacy preserving training of 5 workers with DSSGD (θd = 0.5, θu = 0.5, γ = 10, τ = 0.0001; learning rate: 0.9) |
| D | Local training without collaboration of 5 workers using stochastic gradient descent |
aDSSGD: distributed selective stochastic gradient descent.
Figure 6Performances achieved on the test set during the experiments including C_0.1 (left) or C_0.5 (right).
Scores after 200 training epochs.
| Experiment | F1 | ||
|
| Minimum | Mean | Maximum |
| A | — | 0.962 | — |
| B | — | 0.961 | — |
| C_0.1 | 0.955 | 0.955 | 0.956 |
| C_0.5 | 0.694 | 0.840 | 0.944 |
| D | 0.900 | 0.905 | 0.911 |