Literature DB >> 35186118

Approaches to Federated Computing for the Protection of Patient Privacy and Security Using Medical Applications.

Osman Sirajeldeen Ahmed1, Emad Eldin Omer2, Samar Zuhair Alshawwa3, Malik Bader Alazzam4, Reefat Arefin Khan5.   

Abstract

Computing model may train on a distributed dataset using Medical Applications, which is a distributed computing technique. Instead of a centralised server, the model trains on device data. The server then utilizes this model to train a joint model. The aim of this study is that Medical Applications claims no data is transferred, thereby protecting privacy. Botnet assaults are identified through deep autoencoding and decentralised traffic analytics. Rather than enabling data to be transmitted or relocated off the network edge, the problem of the study is in privacy and security in Medical Applications strategies. Computation will be moved to the edge layer to achieve previously centralised outcomes while boosting data security. Study Results in our suggested model detects anomalies with up to 98 percent accuracy utilizing MAC IP and source/destination/IP for training. Our method beats a traditional centrally controlled system in terms of attack detection accuracy.
Copyright © 2022 Osman Sirajeldeen Ahmed et al.

Entities:  

Year:  2022        PMID: 35186118      PMCID: PMC8856810          DOI: 10.1155/2022/1201339

Source DB:  PubMed          Journal:  Appl Bionics Biomech        ISSN: 1176-2322            Impact factor:   1.781


1. Introduction

While Badotra et al. [1] are credited with coining the term “Medical Applications”, the first description of its implementation can be found in [2]. Multiple devices work together to train a shared model in Medical Applications. Multiple clients' parametric improvements are combined over numerous training rounds to achieve this. Several customers compete in each round to improve a globally available model using data that they have access to only locally. Because the models are supposed to be smaller than the dataset, Medical Applications lowers data transmission while simultaneously addressing privacy issues [3]. Also, all calculations may be done on the customers' devices. Maintaining server farms, developing new models, and handling enormous datasets become simpler. While the round-based structure of Medical Applications implies models are smaller than data transmitted, it is feasible that large bandwidth may be required. In particular, mobile customers with restricted data access should have lower communication expenses. Many communication cost-saving strategies have emerged as a consequence. To address privacy issues, Medical Applications has several customers that compete in each round to construct a globally available model using data that they only have access to locally. While the round-based structure of Medical Applications implies models are smaller than data supplied, a large quantity of bandwidth may be required. Customers with restricted data access should save money. In Figure 1, Medical Applications outperforms ex Google researchers that showed next word prediction and emoticon prediction [4] on mobile devices. These examples demonstrate the versatility of Medical Applications.
Figure 1

Working diagram of Medical Applications.

2. Related Work

A Medical Applications approach to developing WIDS models is proposed by Rahman et al. [5]; edge devices can first train their local models using local data. Local models are then averaged to create a global model. Edge devices do not have to share raw training data this way. These devices train a local model and send only model parameters to the server instead of raw training data.

2.1. Privacy and Security

It is critical to differentiate between privacy and security. Even encrypted data may be accessed by intruders. An adversary's principal purpose is to get victim-specific personal data [6]. Privacy is “the control over who receives personal information”. In principle, a privacy-preserving technology would not reveal personal information to anybody other than those who had given authorization. In reality, personal information is only that which can be assigned to a particular individual. Security, on the other hand, is “the absence of danger or threat”. In the contemporary context, this is mostly accidental data dispersion (or data leakage). What happens to the released data is unimportant but potentially dangerous [7-10]. In the world of encryption and cyber security, it is said that no system can be 100% safe. This is because security systems need a key or password. Define a key in limited space. A password, for example, usually has a maximum character count and a limited character set. So, software that verifies all key combinations may be built. This will crack the system given enough computational resources. This is called a brute force attack. In practice, keys of lengths higher than a certain length are commonly regarded as being unbreakable because the number of combinations to check would take far too much time and resources [11]. The costs of trying this would far exceed any value an adversary gets from breaking the security. There is however a balancing acts because performing the encryption and decryption also has an associated computational cost, which makes it a challenge in minimizing security costs while maximizing the security of the system. Another notion of security is security by obscurity, meaning that you can minimize security risks by not disclosing the details of your protective measures [12]. This has been widely criticized and shall be cast aside as a viable security measure for the remainder of the thesis. Any adversary is considered to have complete knowledge of how the system works which includes complete knowledge of the implemented security measures.

3. Background

For security researchers and industry professionals, DL has recently become a hot topic. DNNs, also known as computing (DL), are a subset of AI that are inspired by how the brain works. DL-based architectures may automatically update derived meaning without domain expert knowledge. Feature extraction is a crucial aspect of feature engineering [13]. Feature extraction affects classification performance. The NN ML approach appeared in the 1950s. It can automatically extract and categorise characteristics without human intervention. The classical NN performs beautifully. Using sophisticated NN (deep learning), however, the feature engineering step may be totally avoided [14-16]. Figure 2 shows the training and testing procedures in classic ML and DL algorithms. As a result, the DL outperformed other long-standing AI applications in several sectors.
Figure 2

Computing-based classification process.

Figure 2 depicts DL architecture classifications. The terms neural networks (NNs), computing (ML), and computing (DL) are all intertwined in AI discourse. All of these fields are frequently misunderstood. DL is a branch of computing that developed from neural networks (NNs). By processing data and generating patterns, this simulates the workings of the human brain. When it comes to DL, the most important part is the NNs, and the term “many NNs” usually means just that: many NNs [17]. Vanishing and exploding gradients and, most importantly, the lack of high-performance computing systems arise when NNs are deep. New DL architectures have emerged, as have advancements in optimizers, activation functions, loss functions, and the vanishing and ballooning gradient concerns. There are two kinds of DL architecture: generative and discriminative [18]. Ideas are generated using deep Boltzmann machines, deep autoencoders, deep belief networks, and recurrent structures. This is done using recurrent structures and convolutional neural networks (CNN).

3.1. Proposed Architecture

A decentralised FL-based deep anomaly detection engine detects botnet attacks. These security gateways operate and monitor traffic to and from Medical Applications equipment. Forensic-based Medical Applications security gateways include FL and anomaly detection. Port mirroring maintains track of network activity. The network traffic entering and leaving the Medical Applications security gateway is monitored for botnets. Infected computing devices typically send messages in unexpected directions [19]. The Medical Applications security gates may communicate after linked to its FL server, which hosts device models. The security gateway will ask the FL server for the deep autoencoder model. Based on data from the worldwide FL server, which shares the same network as the proposed Medical Applications device, the FL model is exclusive for computers. As previously stated, the security gateway hardware may be configured to operate with a variety of devices. We claim a security gateway may employ port mirroring to capture and process network traffic. We may provide information to our virtual worker by snapping a picture. These portals may also accommodate many virtual workers. Unlimited virtual employees may be hosted by each security gateway. It is up to employees to obtain the company a specific gadget. The many to one link between virtual works and security gates. Although a gateway can handle several virtual employees, only one is required per gateway. In this thesis, we will regard security gateways and virtual employees as one-to-one interactions.

3.2. Deep Autoencoder

Half of the network encodes and decodes. Autoencoders are a subset of neural networks. PCA and PCA are closely related, but PCA is much more flexible. Unlike PCA, which can only perform linear transformations, autoencoders can encode data in nonlinear ways [20]. Using autoencoders can maximize data utilization by reducing reconstruction error. Each layer has the same number of neurons using autoencoders (input and output). This is done in the first step of the deep autoencoder, which uses PyTorch linear layers for all steps of the ML process, encoding and decoding continuously as each layer is added and subtracted. Data from the first layer represents the source IP, destination IP, and UDP/TCP socket details and is encoded to 75% of its original size before being sent to the second layer for decompression The input from the previous layer will be passed on to the next subsequential layer for encoding. Half of the input size will be encoded in the next encoding layer, reducing the size by 50%. The input is reduced by 33% in size in the third layer, which continues the encoding process. The input will be encoded down to 25% of the previous step again in the final encoding layer. The compression level is the lowest at this point The effects of the encoding stage will be undone during the decoding stage. It decodes the input and then adds on to the size of it for the next layer. Using the same encoding and decoding values, the same decompression aids as input features. Using the decoder's opposing direction helps to produce a decompressed dataset that is not 1-to-1 identical to the input, as well as expanding and zeroing out some data points to help produce the threshold The output layer will recreate the encoding and decoding process. After encoding and decoding the network traffic's behaviours, a threshold is generated and used for testing by comparing the input and output

3.2.1. Massively Distributed

Because data is stored across a large number of clients, the amount of data available to each client may be significantly smaller than the average amount of data available to each client.

3.2.2. Non-IID

When compared to other clients, the data provided for a particular client may be taken from dramatically different distributions. This means that the data that is readily available in the local area does not accurately reflect the broader data dispersion.

3.2.3. Example

The photographs stored on a cat enthusiast's mobile phone may be radically different from those kept on a vehicle enthusiast's mobile phone.

3.2.4. Unbalanced

The amount of data that is available for a single customer can vary significantly from one client to the next. The centralised model is the most widely used computing technique for decentralised data since it is the most conventional. Because it is explained, it is possible to see how this model differs from that of collaborative techniques in practice.

3.3. Matrix

The following metrics are used to determine overall performance of the IDS model: Detection accuracy: How many samples were correct out of the total sample population? (ii) Recall: Fraction of relevant instances over the total amount of relevant instances. (iii) F1 score: Weighted average of the precision and recall. (iv) False positive rate: The rate at which alerts are generated for normal samples. (v) False negative rate: The rate at which attacks are missed.

4. Results and Discussion

Traditional processes are compared to the suggested federated computing method. Inspiring outcome of the data collection is being used to assess the method's efficacy. First, the new approach's detection performance is compared to the old ones. The results of the tests are shown in Tables 1, 2, and 3 and Figures 3 and 4. The suggested strategy outperforms existing methods in terms of detection rate. The suggested detection method is assessed for precision, recall, accuracy, detection time, false positives, and memory utilization. Table 2 clearly shows that the new strategy outperforms the old.
Table 1

Parameter analysis table for the proposed federated computing method.

ParametersProposed federated computing method
Precision (%)94
Recall (%)92
Accuracy (%)99.9
Detection time (s)32
False positive (%)0.77
Memory utilization (MB)11
Table 2

Parameters.

ParametersConventional methodsProposed federated computing methodImprovement rate
Precision (%)85946.38
Recall (%)83927.61
Accuracy (%)95.3499.95.91
Detection time (s)653233
False positive (%)1.80.771.03
Memory utilization (MB)301119
Table 3

Matrix comparison table between the existing and proposed method.

MetricProposed methodExisting method
False positive (%)1.62.9
Detection rate known (%)99.899.1
Unknown (%)6730.5
Figure 3

Comparison graph between the existing and proposed methods.

Figure 4

Comparison graph.

Figure 5 shows the TN, TP, FP, and FN rates for studies with input dimensions ranging from 15 to 115. These matrices represent the non-FL baseline and the proposed FL techniques' lowest and highest tested input features.
Figure 5

Parameter analysis graph for the proposed method.

Figure 3 displays false positives that resulted in results up to 43954 on the non-FL figure; contrast this with which shows the same parameters but using a multiworker technique. This is a positive reflection on the model, which maintains performance even when the number of workers increases. This applies to all input dimensions, including those with larger dimensions than the default. In Table 3, the non-FL model, for example, produced 31 false positives; however, the multiworker model, in Figures 3 and 4 , produced 36 false positives, proving that the model's performance can be maintained across several workers.

5. Conclusion

In this study, we explore how to use Medical Applications datasets in a virtual classroom. Federated and server-trained computing both perform well in terms of wireless intrusion detection. It does not communicate data to a central server, therefore ensuring user privacy. Then, they rush to patch and repair equipment, leaving new and old networks exposed. We built a model for proactive threat detection. FL thrives in business networks, protects computer hardware, and allows complicated computing models. Self-learning gateways provide for edge network threat detection. As shown in the simulation, accuracy and scores are maintained when enough features are available. These devices may protect Medical Applications equipment connected to a corporate network.
  2 in total

1.  Multi-Institutional Deep Learning Modeling Without Sharing Patient Data: A Feasibility Study on Brain Tumor Segmentation.

Authors:  Micah J Sheller; G Anthony Reina; Brandon Edwards; Jason Martin; Spyridon Bakas
Journal:  Brainlesion       Date:  2019-01-26

2.  Patient clustering improves efficiency of federated machine learning to predict mortality and hospital stay time using distributed electronic medical records.

Authors:  Li Huang; Andrew L Shea; Huining Qian; Aditya Masurkar; Hao Deng; Dianbo Liu
Journal:  J Biomed Inform       Date:  2019-09-24       Impact factor: 6.317

  2 in total
  4 in total

1.  Assessment of the Knowledge Level of First Aid among Medical Students in Work Environment.

Authors:  Asma Alanazi; Saad Algethami; Mohammed Alnafisah; Saoud Alhussayen; Fares Almutairi; Sultan Almureef; Maram Albalawi; Farida Habib
Journal:  J Environ Public Health       Date:  2022-05-30

2.  Identification Level of Awareness and Knowledge of Emirati Men about HPV.

Authors:  Suzan Al Shdefat; Shamsa Al Awar; Nawal Osman; Howaida Khair; Gehan Sallam; Sara Maki
Journal:  J Healthc Eng       Date:  2022-03-21       Impact factor: 2.682

3.  Diagnosis of Breast Cancer Pathology on the Wisconsin Dataset with the Help of Data Mining Classification and Clustering Techniques.

Authors:  Walid Theib Mohammad; Ronza Teete; Heyam Al-Aaraj; Yousef Saleh Yousef Rubbai; Majd Mowafaq Arabyat
Journal:  Appl Bionics Biomech       Date:  2022-04-01       Impact factor: 1.781

4.  Computed Tomography (Ct) Scan Assisted Machine Learning in the Management of Artifacts Related to Paranasal Sinuses and Anterior Cranial Fossa.

Authors:  Abdullah Musleh
Journal:  Comput Intell Neurosci       Date:  2022-10-07
  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.