| Literature DB >> 34696003 |
Yunji Yang1, Yonggi Hong1, Jaehyun Park1.
Abstract
In this paper, efficient gradient updating strategies are developed for the federated learning when distributed clients are connected to the server via a wireless backhaul link. Specifically, a common convolutional neural network (CNN) module is shared for all the distributed clients and it is trained through the federated learning over wireless backhaul connected to the main server. However, during the training phase, local gradients need to be transferred from multiple clients to the server over wireless backhaul link and can be distorted due to wireless channel fading. To overcome it, an efficient gradient updating method is proposed, in which the gradients are combined such that the effective SNR is maximized at the server. In addition, when the backhaul links for all clients have small channel gain simultaneously, the server may have severely distorted gradient vectors. Accordingly, we also propose a binary gradient updating strategy based on thresholding in which the round associated with all channels having small channel gains is excluded from federated learning. Because each client has limited transmission power, it is effective to allocate more power on the channel slots carrying specific important information, rather than allocating power equally to all channel resources (equivalently, slots). Accordingly, we also propose an adaptive power allocation method, in which each client allocates its transmit power proportionally to the magnitude of the gradient information. This is because, when training a deep learning model, the gradient elements with large values imply the large change of weight to decrease the loss function.Entities:
Keywords: aggregated gradient updating; federated learning; image classification; wireless backhaul
Year: 2021 PMID: 34696003 PMCID: PMC8537050 DOI: 10.3390/s21206791
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Federated Learning over wireless communication system for target classification.
Figure 2CNN module for handwriting character recognition.
The values for the hyperparameters of the CNN module for handwriting character recognition.
| Values | |
|---|---|
| The number of layers | 3 |
| The number of filters at each layer | 3 |
| Filter size (The 1st layer), |
|
| Filter size (The 2nd layer), |
|
| Filter size (The 3rd layer), |
|
| Optimizer | ADAM optimizer [ |
| Learning rate, | 0.001 |
Figure 3(a) Classification accuracy and (b) CE loss curves at dB.
Figure 4(a) Classification accuracy and (b) CE loss curves at dB.
Figure 5Classification accuracy according to various threshold levels when the updating method with MRC and thresholding in Section 4.2.2 are exploited for (a) dB and (b) dB.
Figure 6Comparison of classification accuracy with (a) and (b) for dB.
Confusion matrix for hand writing character recognition of the proposed gradient updating method.
| Predicted | True Label | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Label | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 |
| 0 |
| 0.001 | 0.020 | 0.004 | 0.005 | 0.006 | 0.008 | 0.003 | 0.022 | 0.007 |
| 1 | 0 |
| 0.002 | 0 | 0.005 | 0.001 | 0.002 | 0.002 | 0.001 | 0.001 |
| 2 | 0 | 0.004 |
| 0.005 | 0 | 0 | 0 | 0.015 | 0.004 | 0 |
| 3 | 0 | 0.001 | 0.006 |
| 0 | 0.006 | 0 | 0 | 0.007 | 0.007 |
| 4 | 0 | 0.001 | 0.001 | 0 |
| 0 | 0.002 | 0.001 | 0.005 | 0.004 |
| 5 | 0 | 0.001 | 0 | 0.019 | 0.001 |
| 0.005 | 0.001 | 0.003 | 0 |
| 6 | 0.010 | 0.005 | 0.003 | 0 | 0.016 | 0.015 |
| 0 | 0.007 | 0.001 |
| 7 | 0.001 | 0.001 | 0.016 | 0.007 | 0.001 | 0.003 | 0 |
| 0.007 | 0.016 |
| 8 | 0 | 0.007 | 0.009 | 0.002 | 0.001 | 0.011 | 0.001 | 0.003 |
| 0.006 |
| 9 | 0.005 | 0 | 0.001 | 0 | 0.042 | 0.003 | 0.001 | 0.026 | 0.029 |
|
Confusion matrix for hand writing character recognition of equal-weight combining based gradient updating method.
| Predicted | True Label | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Label | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 |
| 0 |
| 0.004 | 0.012 | 0.004 | 0.011 | 0.026 | 0.010 | 0.002 | 0.005 | 0.001 |
| 1 | 0 |
| 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
| 2 | 0.977 | 0.875 |
| 0.877 | 0.932 | 0.777 | 0.971 | 0.842 | 0.919 | 0.943 |
| 3 | 0 | 0 | 0 |
| 0 | 0 | 0 | 0 | 0 | 0 |
| 4 | 0.009 | 0.085 | 0.026 | 0.113 |
| 0.172 | 0.017 | 0.126 | 0.059 | 0.036 |
| 5 | 0 | 0 | 0 | 0 | 0 |
| 0 | 0 | 0 | 0 |
| 6 | 0 | 0 | 0 | 0 | 0 | 0 |
| 0 | 0 | 0 |
| 7 | 0 | 0.025 | 0.001 | 0.002 | 0.016 | 0 | 0 |
| 0 | 0.020 |
| 8 | 0.004 | 0.011 | 0 | 0.004 | 0.002 | 0.026 | 0.002 | 0.007 |
| 0.001 |
| 9 | 0 | 0.001 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
|