| Literature DB >> 35755701 |
Aiguo Chen1, Yang Fu1, Zexin Sha1, Guoming Lu1.
Abstract
Federated learning is a distributed machine learning framework that enables distributed nodes with computation and storage capabilities to train a global model while keeping distributed-stored data locally. This process can promote the efficiency of modeling while preserving data privacy. Therefore, federated learning can be widely applied in distributed conjoint analysis scenarios, such as smart plant protection systems, in which widely networked IoT devices are used to monitor the critical data of plant production to improve crop production. However, the data collected by different IoT devices can be dependent and identically distributed (non-IID), causing the challenge of statistical heterogeneity. Studies have also shown that statistical heterogeneity can lead to a marked decline in the efficiency of federated learning, making it challenging to apply in practice. To promote the efficiency of federated learning in statistical heterogeneity scenarios, an adaptive client selection algorithm for federated learning in statistical heterogeneous scenarios called ACSFed is proposed in this paper. ACSFed can dynamically calculate the possibility of clients being selected to train the model for each communication round based on their local statistical heterogeneity and previous training performance instead of randomly selected clients, and clients with heavier statistical heterogeneity or bad training performance would be more likely selected to participate in the later training. This client selection strategy can enable the federated model to learn the global statistical knowledge faster and thereby promote the convergence of the federated model. Multiple experiments on public benchmark datasets demonstrate these improvements in the efficiency of the models in heterogeneous settings.Entities:
Keywords: adaptive client selection; distributed conjoint analysis; federated learning; machine learning; statistical heterogeneity
Year: 2022 PMID: 35755701 PMCID: PMC9218865 DOI: 10.3389/fpls.2022.908814
Source DB: PubMed Journal: Front Plant Sci ISSN: 1664-462X Impact factor: 6.627
Adaptive client selection enabling federated learning (ACSFed). K clients are selected from N clients with fraction c; η is the learning rate, P is the probability matrix; H is the cumulative model strength matrix.
| |
| Initialize model ω0, |
| |
| |
| if t=0: |
| client_set = {randomly selected |
| else: |
| client_set = {select |
| |
| Transmit ω |
| Receive |
| |
| |
| |
| |
| |
| Receive model ω |
| Calculate |
| Initialize loss list |
| Θ← {split local data into batches with size |
| |
| |
| Add training loss to |
| ω |
| |
| |
| |
| Calculate |
| Transmit |
Details of the dataset.
|
|
|
|
|
|---|---|---|---|
|
| |||
| MNIST | 10 | 28 × 28 × 1 | Hand-writing number image |
| Fashion MNIST | 10 | 28 × 28 × 1 | Wearing image |
| CIFAR 10 | 10 | 32 × 32 × 3 | Common things image |
Figure 1Training loss on MNIST in 2-class non-IID scenario.
Figure 2Test accuracy on MNIST in 2-class non-IID scenario.
Figure 3Training loss on MNIST in the 1-class non-IID scenario.
Figure 4Test accuracy on MNIST in the 1-class non-IID scenario.
Figure 5Training loss on fashion MNIST in 2-class non-IID scenario.
Figure 6Test accuracy on fashion MNIST in 2-class non-IID scenario.
Figure 7Training loss on fashion MNIST in the 1-class non-IID scenario.
Figure 8Test accuracy on fashion MNIST in the 1-class non-IID scenario.
Figure 9Test accuracy on CIFAR 10 in the 1-class non-IID scenario.
Figure 10Comparison results between ACSFed and FedProx on fashion MNIST.
Training performance of ACSFed and FedAvg.
|
|
|
|
|
|---|---|---|---|
| MNIST | ACSFed | 3/25 | 6/72 |
| FedAvg | 5/27 | 9/NULL | |
| Fashion MNIST | ACSFed | 5/108 | 32/946 |
| FedAvg | 8/107 | 35/NULL | |
|
|
| ||
| CIFAR 10 | ACSFed | 12/381 | 235/NULL |
| FedAvg | 20/674 | 927/NULL |