| Literature DB >> 35458968 |
Evgenia Novikova1, Dmitry Fomichov1, Ivan Kholod1, Evgeny Filippov2.
Abstract
Wearable devices and smartphones that are used to monitor the activity and the state of the driver collect a lot of sensitive data such as audio, video, location and even health data. The analysis and processing of such data require observing the strict legal requirements for personal data security and privacy. The federated learning (FL) computation paradigm has been proposed as a privacy-preserving computational model that allows securing the privacy of the data owner. However, it still has no formal proof of privacy guarantees, and recent research showed that the attacks targeted both the model integrity and privacy of the data owners could be performed at all stages of the FL process. This paper focuses on the analysis of the privacy-preserving techniques adopted for FL and presents a comparative review and analysis of their implementations in the open-source FL frameworks. The authors evaluated their impact on the overall training process in terms of global model accuracy, training time and network traffic generated during the training process in order to assess their applicability to driver's state and behaviour monitoring. As the usage scenario, the authors considered the case of the driver's activity monitoring using the data from smartphone sensors. The experiments showed that the current implementation of the privacy-preserving techniques in open-source FL frameworks limits the practical application of FL to cross-silo settings.Entities:
Keywords: differential privacy; driver activity monitoring; federated learning; homomorphic encryption; open-source federated learning frameworks; privacy; secure multi-party computations
Mesh:
Year: 2022 PMID: 35458968 PMCID: PMC9029817 DOI: 10.3390/s22082983
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1FL components and attack entry points during the training phase.
Figure 2Overview of privacy-preserving mechanisms adopted for FL.
Properties of privacy mechanisms adopted to FL settings.
| Privacy Mechanism | Privacy of Inputs | Privacy of Computations | Security Model | Dropouts | Support of Vertically Partitioned Data * | Accuracy | Specific Features and Requirements |
|---|---|---|---|---|---|---|---|
| Local DP | + | − | Honest-but-curious aggregating server | + | + ** | Decreases due to noise in comparison to FL models without any privacy-preserving techniques | Requires additional collaboration when setting privacy budget parameters |
| DP with shuffling | + | − | Honest-but-curious aggregating server | + | − | Decreases due to noise | Requires additional entity (shuffler), |
| MPC | + | + | Honest majority | + | − | Comparable to the accuracy of FL models without MPC, | Requires additional computational entities that perform all trusted computations; |
| Secure aggregation | + | − | Honest-but-curious aggregating server | + | − | Similar to the accuracy of FL models without MPC (input data are treated as binary vectors) | Time depends linear on the number of clients and on cryptographic primitives for public key infrastructure |
| HE | + | + | Malicious server Requires a trusted server for the generation of the encryption keys | − | + | Comparable to the accuracy of FL models without MPC, | Extremely high requirements for computational, memory and disc resources of collaborating entities. |
| TEE | + | + | Trusted server | + | − *** | Similar to the accuracy of FL models without any privacy-preserving techniques | Adds light overhead in CPU, memory, energy due to specific computational requirements; |
*—the authors consider the case of vertically partitioned data as a more complicated and not widely used case. It usually requires adoption of additional dataset alignment techniques. **—it requires creating additional data set or application of encryption techniques. ***—this issue is not presented in the literature.
Types of privacy mechanisms implemented in open-source FL systems.
| Framework and Company | TFF 0.17.0 [ | FATE 1.5.0 [ | PFL 1.1.0 [ | FL & DP 0.1.0 [ | FEDn 0.2.0 [ |
|---|---|---|---|---|---|
| DP | ✔ | ✔ | ✔ | ||
| HE | ✔ | ||||
| MPC | ✔ | ✔ | |||
| Secure aggregation | ✔ | ✔ | ✔ | ||
| TEE | ✔ |
Privacy mechanisms implemented in FATE 1.5.0.
| Data Partition Type | Type of Privacy Mechanism | Analysis Model | Implementation of Privacy Mechanism |
|---|---|---|---|
| Horizontal | Data encryption | NN (DNN, RNN, CNN) | Secure Aggregation [ |
| Logistic Regression | Paillier [ | ||
| Vertical | Data encryption | NN (DNN, RNN, CNN) | SecureNN, or Iterative Affine, or Affine Additive HE [ |
| Gradient Boosting Trees | SecureBoost [ | ||
| Linear regression | Paillier [ |
Figure 3Comparison of different parameters for FedAvg and SecAgg strategies implemented in FATE: accuracy (a), training time (b), network traffic (MB) (c).
Figure 4Scheme of secure NN training on vertically partitioned data implemented in the FL framework FATE.
Parameters of NN training using additive HE scheme for different settings.
| Data Set Size | Batch Size | Accuracy | Training Time | Traffic, GB |
|---|---|---|---|---|
| 10,000 | Entire data set | 98.04% | 1:02:03 | 1.71 |
| 10,000 | 1000 | 97.15% | 1:20:15 | 1.72 |
| 10,000 | 100 | 99.58% | 9:30:59 | 1.78 |
Figure 5Comparison of different parameters of training process for Fast SecureBoost strategy with different initial settings in FATE: accuracy (a), training time (b), network traffic (GB) (c).
Linear regression model: experimental settings and results obtained.
| Number of Clients | Data Set Size per Clients (Number of Records) | MSE | Traffic (GB), Arbiter | Training Time |
|---|---|---|---|---|
| 2 | 425,000 | 0.0002 | 11.30 | 7:57:312 |
| 4 | 425,000 | 1.44 | 65.00 | 13:54:18 |
Privacy mechanisms implemented in PFL 1.1.0.
| Data Partition Type | Analysis Model | Type of Privacy Mechanism | Implementation of Privacy Mechanism |
|---|---|---|---|
| Horizontal | NN | Differential privacy | DP-SGD [ |
| Data encryption | Secure Aggregation [ | ||
| Verical | NN | Data encryption | PSI [ |
Figure 6Network traffic (MB) generated by a client when training with FedAvg and SecAgg strategies for different experiment and model settings: 20 rounds (a), 10 rounds (b).
Figure 7Training time of FedAvg and SecAgg for different experiment and model settings: 20 rounds (a), 10 rounds (b).
Figure 8Privacy budget depending on batch size and epoch number.
Figure 9Accuracy of DPSGD strategy for two clients and round number set to 20.
Figure 10Analysis pipeline using MPC protocols implemented in the PFL framework.
Parameters of NN training using MPC for different settings.
| Data Set Size | Batch Size | Training Time | Accuracy | Network Traffic (GB) |
|---|---|---|---|---|
| 10,000 | 10,000 | 0:13:58 | 51.00% | 35.80 |
| 1000 | 0:16:35 | 53.75% | 36.40 | |
| 32 | 2:11:39 | 99.00% | 58.40 | |
| 1,700,000 | 1000 | 27:19:22 | 49.67% | Not measured |
Properties of the privacy mechanisms implemented in FATE and PFL and mapped to the FL system properties.
| Data Partition | Clients’ Settings | Communication Topology | Framework and Privacy Mechanisms | Impact on Overall Learning Process | ||
|---|---|---|---|---|---|---|
| Accuracy | Training Time | Network Traffic | ||||
| Horizontal | Cross-silo | Centralized | FATE and Secure Aggregation | Similar to FedAvg strategy in terms of accuracy. | Similar to the FedAvg strategy. | Similar to the FedAvg strategy. |
| FATE and Homomorphic encryption (Paillier encryption scheme) | Similar to FedAvg strategy. | Extremely time-consuming training strategy. | Extremely high volume of traffic. | |||
| PFL and Secure Aggregation | Comparable to FedAvg strategy. | Comparable to FedAvg strategy. | High network traffic. There is an observable dependency between training time and network performance. | |||
| PFL and Differencial privacy (DP-SGD) | The accuracy is lower the accuracy in FedAvg strategy, but it increases with growth of clients’ number. | High when compared with similar settings without privacy mechanisms. | The network traffic is extremely high and is measured in GB. | |||
| De-centralized | Not supported | Not supported | Not supported | Not supported | ||
| Cross-device | Centralized | Not supported | Not supported | Not supported | Not supported | |
| De-centralized | Not supported | Not supported | Not supported | Not supported | ||
| Vertical | Cross-silo | Centralized | FATE and Homomorphic encryption (SecureNN, SGBDT) | The accuracy is comparable with the training model in centralized mode. | It is a very time-consuming strategy. | It is characterized by a high volume of traffic measured in GB. |
| PFL and MPC (ABY | The accuracy is comparable with training model in centralized mode, however it is significantly lower for neural networks when training on large batches. | It is a very time-consuming strategy. | It is characterized by a high volume of traffic measured in GB. | |||
| De-centralized | Not supported | Not supported | Not supported | Not supported | ||
| Cross-device | Centralized | Not supported | Not supported | Not supported | Not supported | |
| De-centralized | Not supported | Not supported | Not supported | Not supported | ||