| Literature DB >> 34956350 |
Shaashwat Agrawal1, Aditi Chowdhuri1, Sagnik Sarkar1, Ramani Selvanambi1, Thippa Reddy Gadekallu2.
Abstract
Federated learning (FL) is an emerging subdomain of machine learning (ML) in a distributed and heterogeneous setup. It provides efficient training architecture, sufficient data, and privacy-preserving communication for boosting the performance and feasibility of ML algorithms. In this environment, the resultant global model produced by averaging various trained client models is vital. During each round of FL, model parameters are transferred from each client device to the server while the server waits for all models before it can average them. In a realistic scenario, waiting for all clients to communicate their model parameters, where client models are trained on low-power Internet of Things (IoT) devices, can result in a deadlock. In this paper, a novel temporal model averaging algorithm is proposed for asynchronous federated learning (AFL). Our approach uses a dynamic expectation function that computes the number of client models expected in each round and a weighted averaging algorithm for continuous modification of the global model. This ensures that the federated architecture is not stuck in a deadlock all the while increasing the throughput of the server and clients. To implicate the importance of asynchronicity in cybersecurity, the proposed algorithm is tested using NSL-KDD intrusion detection system datasets. The performance accuracy of the global model is about 99.5% on the dataset, outperforming traditional FL models in anomaly detection. In terms of asynchronicity, we get an increased throughput of almost 10.17% for every 30 timesteps.Entities:
Mesh:
Year: 2021 PMID: 34956350 PMCID: PMC8709749 DOI: 10.1155/2021/5844728
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1Architecture of network-based intrusion detection system (NIDS).
Summary of DL-based IDS and FL-based IDS.
| Sl. No. | DL-based IDS | FL-based IDS |
|---|---|---|
| 1 | Constrained to a local system. | Scalable and readily deployable in a network. |
| 2 | Pretrained or active training has limited relationship with similar nodes in the network. | Extensive relationship between models. Promotes general system learning and personalization for each node. |
| 3 | Less effective for evolving intrusions and response to novel attacks. | Structured to learn from every node's individual experience and respond to unforeseen attacks. |
Figure 2Proposed asynchronous federated algorithm.
Algorithm 1Algorithm 1Temporal weighted averaging.
Figure 3Distribution of the types of attacks in the NSL-KDD dataset.
Figure 4Model architecture for the global model and the client models.
Figure 5Number of client models aggregated (top) per timestep and its corresponding time threshold (T0) (bottom).
Figure 6Confusion matrix for the server model after 15 rounds.
Figure 7Comparison of the performance of the federated learning architectures. (a) The plot of the accuracy of the global model in each round. (b) The loss scores of the global model for each round.
Performance comparison of multiple architectures.
| Sl. No. | Algorithm | Clients | Rounds/timesteps | Accuracy |
|---|---|---|---|---|
| 1 | Asynchronous FL with temporal weighted averaging | 30 | 15 | 99.459 |
| 2 | Federated learning | 30 | 15 | 99.11 |
| 3 | FTML (federated teacher mimic learning) [ | 10 | 20 | 98.118 |
| 4 | FSML (federated student mimic learning) [ | 10 | 20 | 98.110 |
| 5 | Collaborative anomaly detection [ | 4 | 4 | 98.24 |
| 6 | PHEC in federated setup [ | 4 | Adaptive | 88.42 |