| Literature DB >> 35136708 |
Amine Boulemtafes1, Abdelouahid Derhab2, Yacine Challal3.
Abstract
In recent years, deep learning in healthcare applications has attracted considerable attention from research community. They are deployed on powerful cloud infrastructures to process big health data. However, privacy issue arises when sensitive data are offloaded to the remote cloud. In this paper, we focus on pervasive health monitoring applications that allow anywhere and anytime monitoring of patients, such as heart diseases diagnosis, sleep apnea detection, and more recently, early detection of Covid-19. As pervasive health monitoring applications generally operate on constrained client-side environment, it is important to take into consideration these constraints when designing privacy-preserving solutions. This paper aims therefore to review the adequacy of existing privacy-preserving solutions for deep learning in pervasive health monitoring environment. To this end, we identify the privacy-preserving learning scenarios and their corresponding tasks and requirements. Furthermore, we define the evaluation criteria of the reviewed solutions, we discuss them, and highlight open issues for future research.Entities:
Keywords: Deep learning; Deep neural network; Pervasive health monitoring; Privacy; e-Health; m-Health
Year: 2022 PMID: 35136708 PMCID: PMC8813181 DOI: 10.1007/s12553-022-00640-3
Source DB: PubMed Journal: Health Technol (Berl) ISSN: 2190-7196
Fig. 1DL-based PHM components and process flow
Evaluation criteria for PHM environment adequacy
| Effectiveness | [ | ||
| Efficiency | [ [ | ||
| Privacy | [ | [ | [ |
[]*: desired properties
Privacy-preserving solutions for training a local model
| Key concept | Ref | Main characteristics |
|---|---|---|
| Homomorphic Encryption | Q. Zhang et al. [ | Fully BGV HE | Taylor theorem polynomial approximation |
| Bu et al. [ | Fully BGV HE | Maclaulin formula polynomial approximation | |
| Phong et al. [ | Partially additive LWE-based and Paillier HE | TLS/SSL secure channels | |
| X. Zhang et al. [ | Partially lightweight El Gamal HE | Shamir’s threshold secret sharing | Local differential privacy | |
| Hao et al. [ | Partially additive HE | Differential privacy | |
| Partial sharing | Shokri and Shmatikov [ | Partial sharing of parameters | Laplace differential privacy | Sparse vector technique |
| Liu et al. [ | Partial sharing of parameters | |
| Transformation | Zhao et al. [ | Functional exponential mechanism | Polynomial approximation for objective function | Cryptography and hashing against eavesdrop attacks |
| Hartmann and West [ | Cancelable noise (differential privacy) | Anonymization network (such as Tor) | |
| Fu et al. [ | Mixup data augmentation | |
| Shared model | Servia-Rodriguez et al. [ | Start from the weights and bias of the shared model | (Optional) Differential privacy for training the shared model |
Solutions for training a local model vs evaluation criteria
| [ | None(7) | TD [and weights] encrypted | weights encrypted | AFR: | ||
| [ | None(7) | TD [and weights] encrypted | weights encrypted | AFR: | ||
| [ | SAN | None(8) | TD not shared [and local shared weights encrypted] | not shared | ADV: | |
| [ | CAN | TD not shared [and shared gradients perturbed and encrypted, | not shared | - | ||
| [ | Quite high | None(8) | TD not shared [and shared gradients perturbed and encrypted, | not shared | ADV: tolerate collusion of server with multiple users | |
| [ | CAN | None(8) | TD not shared [and only fraction of local parameters shared perturbed, | not shared | ADV: | |
| [ | RR ~ 90% | Round-robin: Asynchronous: None(8) | TD not shared [and only fraction of local parameters shared, | not shared, | ||
| [ | MRE High | None(8) | TD not shared [and objective functions (thus gradients) perturbed, | ADV: active & passive participants AFR/MOR: | ||
| [ | N/A | TD not shared [and local gradients perturbed with cancelable noise] | ADV: malicious server, honest participants > = 2 | |||
| [ | Up to High | [None(8) | | TD not shared [and local parameters obtained from mixup input data, | - | ||
| [ | Up to High | None | TD not shared + resistant to model inversion | - | ||
ind Individual training, col Collaborative Accuracy training mainly on the basis of authors evaluations, SAN can reach Same As Non-private model, CAN can reach Close to/Almost Non-private model, RR Reconstruction Rate, MRE Mean Relative Error, OVC overhead is relative to the non-private model, DIT [ col] high if the training round stops for all participants, none, if the training round is not affected, [ ind] high if the participant training process stops, none if the process does not stop, ADV Adversary model, AFR Activation Function Restrictions, MOR Model Restrictions, APX Approximated
(1) Due to polynomial approximation.
(2) Trade-off with privacy
(3) Provided that a shared model can be trained at the cloud, and that clients have enough samples to personalize their local models
(4) Due to Homomorphic encryption.
(5) Due to iterative interaction between client and server
(6) Due to local training
(7)Training round only needs local data to be transferred
(8) However, user’s local training is not considered in the global model until transferred
(9) Depends on the server round policy, i.e., wait for late users? Indefinitely, or for a certain period, …etc.
(10) Trade-off with accuracy
(11) Global gradients are not protected
(12) Structure shared
* Privacy guarantees and limitations relative to indirect leakage are distinguished between square brackets []
Comparison of key concepts for training a local model
| Key concept | Learning | Accuracy | Client overhead | Dropout impact | Data privacy | Local model privacy |
|---|---|---|---|---|---|---|
| HE | IND—COL | |||||
| PS | COL | |||||
| TRA | COL | |||||
| SM | IND |
HE Homomorphic encryption, PS Partial sharing, TRA Transformation, SM Shared model, IND Individual learning COL Collaborative learning, Dropout impact—COL high, if the training round stops for all participants; none, if the training round is not affected, IND high, if the participant training process stops, none, if the process does not stop, trade-off between accuracy and privacy
a in case the coordination or a threshold of participants and/or their transmitted information are required in the process at each round
b in case of using perturbation against inter-participants protection
Privacy-preserving solutions for training a remote model
| Key concept | Ref | Main characteristics |
|---|---|---|
| Homomorphic Encryption | Q. Zhang et al. [ | Partially Paillier HE | Outsourcing non-linear computations to the client |
| Partial sharing | Shokri and Shmatikov [ | Partial sharing of parameters | Laplace differential privacy | Sparse vector technique |
| Liu et al. [ | Partial sharing of parameters | |
| Transformation | Lyu et al. [ | Repeated Gompertz (RG) for data perturbation | Row-orthogonal random projection (RP) matrix for projecting high-dimensional data to lower dimension |
| Zhao et al. [ | Functional exponential mechanism | Polynomial approximation for objective function | Cryptography and hashing against eavesdrop attacks | |
| Hartmann and West [ | Cancelable noise (differential privacy) | Anonymization network (such as Tor) | |
| Fu et al. [ | Mixup data augmentation | |
| Model splitting | Yu et al. [ | 1st convolutional layer on local | Step-wise activation functions | CNN |
| Abuadbba et al. [ | Part of layers on local | differential privacy | |
| Dong et al. [ | 1st layer on local | Dropping connections and activation outputs | Dropout and Dropconnect |
Solutions for training a remote model vs evaluation criteria
| [ | WLS(1) | TD [and intermediate results] encrypted | Only | AFR/MOR: tested on DNN and CNN | ||
| [ | CAN | None(7) | TD not shared [and only fraction of local parameters shared perturbed, | ADV: | ||
| [ | Round-robin: Asynchronous: None(7) | TD not shared [and only fraction of local parameters shared, | ||||
| [ | Low communication | None(6) | TD perturbed and projected to lower dimension, | not shared | - | |
| [ | MRE High | None(7) | TD not shared [and objective functions (thus gradients) perturbed, | ADV: active & passive participants AFR/MOR: | ||
| [ | N/A | TD not shared [and local gradients perturbed with cancelable noise] | ADV: malicious server, | |||
| [ | Up to High | [None(7)| | TD not shared [and local parameters obtained from mixup input data, | - | ||
| [ | Low | None(10) | TD not shared [and local output perturbed through step-wise local activation function, | Only | ADV: supports malicious attacks AFR: | |
| [ | - With differential privacy (DP): - No DP: WLS | - With DP: low - No DP: | None(10) | - With DP: TD not shared, [and local output perturbed, - No DP: TD not shared, | - No DP: | MOR: 1-dimension CNN |
| [ | N/A | Low | None(10) | Local outputs protected by droppings | Only | - |
Accuracy mainly on the basis of authors evaluations.
WLS Without Loss, CAN can reach Close to/Almost Non-private model, RR Reconstruction Rate, MRE Mean Relative Error, OVC overhead is relative to the non-private model, DIT high, if the training round stops for all participants, low if the training round stops for only the dropped participant, none if the training round is not affected, ADV Adversary model, AFR Activation Function Restrictions, MOR Model Restrictions
(1) No approximation is involved
(2) Trade-off with privacy
(3) Due to Homomorphic encryption
(4) Due to iterative interaction between client and server
(5) Due to local training
(6) Training round only needs local data to be transferred
(7) but user’s local training not considered in the global model until transferred
(8) Depends on the server round policy: wait for late users? Indefinitely, for a certain period, …
(9) Trade-off with accuracy
(10) After the local partition has been executed, and local output transferred
(11) If local output has not been transferred
(12) Trade-off between the privacy of the model an the privacy of data
* Privacy guarantees and limitations relative to indirect leakage are distinguished between square brackets []
Comparison of key concepts for training a remote model
| Key concept | Accuracy | Client overhead | Dropout impact | Data privacy | Remote model privacy |
|---|---|---|---|---|---|
| HE | |||||
| PS | |||||
| TRA | |||||
| MS |
HE Homomorphic encryption, PS Partial sharing, TRA Transformation, MS Model splitting,Dropout impact—high, if the training round stops for all participants, low if it stops only for the dropped participant, none if it is not affected, trade-off: between accuracy and privacy
a in case the coordination or a threshold of participants and/or their transmitted information are required in the process at each round
b in case of distributed (federated) learning
c in case perturbation is cancelable
d if the local output has not been transferred, otherwise none
Privacy-preserving solutions for remote inference
| Key concept | Ref | Main characteristics |
|---|---|---|
| Homomorphic Encryption | Gilad-Bachrach et al. [ | Leveled YASHE HE | Polynomial approximation of activation function |
| Baryalai et al. [ | Partially Paillier HE | Non-colluding dual clouds | Diffie-Hellman key exchange | Random salt | Classification | |
| Chabanne et al. [ | Fully BGV HE | Low degree polynomial approximation of activation function | Batch normalization | Classification | CNN with depth > 2 | |
| Hesamifard et al. [ | Leveled HE | Polynomial approximation: derivative of ReLU based approach and Sigmoid, Tanh, over a symmetric interval | CNN | |
| Zhu and Lv [ | Partially Paillier HE | Interactive protocol between client and server for ReLU computation | |
| Vizitiu et al. [ | Fully MORE HE | |
| SMC | Huang et al. [ | Additive secret-sharing | Secure computations | CNN feature extractor | Non-colluding dual edge servers |
| Ma et al. [ | Secret sharing | Partially El Gamal HE | Low-degree polynomial approximation of activation function | Non-colluding dual servers | |
| Li et al. [ | Secret sharing | Triplet generation | Fully YASHE HE | Two non-colluding servers | Asynchronous computation | Garbled circuits | CNN | |
| Transformation | Leroux et al. [ | Generative Adversarial Networks | Neural-network-based obfuscation |
| Raval et al. [ | Generative Adversarial Networks | Neural-network-based obfuscation | |
| Xu et al. [ | Neural-network-based obfuscation | |
| Model splitting | Osia et al. [ | Feature extractor on local | Siamese architecture | Dimensionality reduction: PCA and auto-encoder | Symmetric gaussian noise |
| Chi et al. [ | Bipartite model | Interactive adversarial deep networks | |
| Yu et al. [ | 1st convolutional layer on local | Step-wise activation functions | CNN | |
| Dong et al. [ | 1st layer on local | Dropping connections and activation outputs | Dropout and Dropconnect |
Remote inference solutions vs evaluation criteria
| Ref | Effectiveness | Efficiency | Privacy guarantees and limitations | Notes | |||
|---|---|---|---|---|---|---|---|
| [ | Up to 99% | None | encrypted | encrypted | not shared | AFR: | |
| [ | N/A | None | IN encrypted, [ | Encrypted, masked with random salt | not shared | AFR: - | |
| [ | CTN | None | encrypted | encrypted | not shared | AFR: MOR: focus on CNN | |
| [ | CTN | None | | encrypted | encrypted | not shared | AFR: MOR: focus on CNN | |
| [ | CTN | encrypted | encrypted | not shared | AFR: MOR: focus on CNN | ||
| [ | Up to ITN | None | encrypted | encrypted | not shared | ||
| [ | WLS | Low | None | encrypted: split into two shares | encrypted into two shares | ADV: AFR: focus on ReLU MOR: | |
| [ | N/A | None | encrypted | encrypted into two shares | AFR: - | ||
| [ | Preserves high accuracy | Low | None | encrypted: split into two shares | encrypted into two shares | AFR/MOR: focus on CNN - | |
| [ | None | obfuscated | - obfuscator and deobfuscator trained competitively | ||||
| [ | None | obfuscated | not shared | - obfuscator and deobfuscator trained competitively, and with the help of of the main classifier | |||
| [ | None | obfuscated | not shared | ADV: edge devices AFR: - o | |||
| [ | Acceptable (according to authors) | None(7) | | IN not shared, [and fine-tuned features are output instead, | MOR: focus on CNN, - | |||
| [ | Up to high | None(7) | | IN not shared, [and reversibility of local output strengthened, | ADV: adversary can have access to the remote party, and intermediate states - Bipartite model needs to be trained concurrently with the defender | |||
| [ | Up to good | Low | None(7) | | IN not shared, [and local output from step-wise activation function, | Only | ADV: supports malicious attacks AFR: | |
| [ | No noticeable loss | Low | None(7) | | IN not shared, [and local output protected by a dropping strategy, | Only | - | |
Accuracy mainly on the basis of authors evaluations
WLS Without Loss, CTN Close To Non-private model, ITN Identical To Non-private model, OVC overhead is relative to the non-private model, DOB Depends on the obfuscator network and its output, DLP Depends on the local partition, DII high, if the inference process stops, none if the process does not stop, ADV Adversary model, AFR Activation Function Restrictions, MOR Model Restrictions, APX Approximated
(1) Due to polynomial approximation of activation functions
(2) Tested using only non-heavyweight inference models
(3) Different obfnets may achieve different accuracy results, the effectiveness can be non-stable as it may differ from user to user or from an obfnet to another for a same user
(4) Trade-off with privacy
(5) Due to Homomorphic encryption
(6) Due to iterative interaction between client and server
(7) After the local partition has been executed, and local output transferred
(8) If the local output has not been transferred
(9) Trade-off with accuracy
(10) Trade-off between the privacy of the model an the privacy of data
(11) When addressing noise growth
* Privacy guarantees and limitations relative to indirect leakage are distinguished between square brackets []
Comparison of key concepts for remote inference
| Key concept | Accuracy | Client overhead | Dropout impact | Data privacy | Inference privacy | Model privacy |
|---|---|---|---|---|---|---|
| HE | ||||||
| SMC | ||||||
| TRA | ||||||
| MS |
Dropout impact - high, if the inference process stops; none, if the process does not stop. trade-off: between accuracy and privacy
HE Homomorphic encryption, SMC Secure Multiparty Computation, TRA Transformation, MS Model splitting
a if activation function are approximated and depending on the polynomial approximation use
b if homomorphic encryption is used
c in case refreshing noise and/or the computation of activation functions are performed by the client
d if the local output has not been transferred, otherwise non
e in case of obfuscators that need during their training to back-propagate through the main mode
f in order to compute activation functions, Intermediate Results (IR) are shared with the introduced 2nd server without protection