| Literature DB >> 36244993 |
Wentao Li1, Jiayi Tong2, Md Monowar Anjum3, Noman Mohammed3, Yong Chen2, Xiaoqian Jiang4.
Abstract
OBJECTIVES: This paper developed federated solutions based on two approximation algorithms to achieve federated generalized linear mixed effect models (GLMM). The paper also proposed a solution for numerical errors and singularity issues. And showed the two proposed methods can perform well in revealing the significance of parameter in distributed datasets, comparing to a centralized GLMM algorithm from R package ('lme4') as the baseline model.Entities:
Keywords: Federated learning; GLMM; Gauss–Hermite approximation; Laplace approximation; Mixed effects
Mesh:
Year: 2022 PMID: 36244993 PMCID: PMC9569919 DOI: 10.1186/s12911-022-02014-1
Source DB: PubMed Journal: BMC Med Inform Decis Mak ISSN: 1472-6947 Impact factor: 3.298
Fig. 1Schema of federated learning model in multiple geographically distributed healthcare institutions. The local institutions periodically exchange intermediate statistics and update the convergence situation of the global model
The summary of data in each setting
| Setting | Number of sites | Sample size in eachsite | Variance |
|---|---|---|---|
| 1 | 2 | 500 | Small |
| 2 | 2 | 500 | Large |
| 3 | 10 | 500 | Small |
| 4 | 10 | 500 | Large |
| 5 | 2 | 30 | Small |
| 6 | 2 | 30 | Large |
| 7 | 10 | 30 | Small |
| 8 | 10 | 30 | Large |
Fig. 2The difference from coefficients to the true parameters that are used to generate data. (Left) The distributed GLMM with Laplace approximation; (Middle) The distributed GLMM with 2-degree Gauss–Hermite approximation. Reminds that is the intercept; (Right) The benchmark of centralized GLMM in R package
Fig. 3The comparison of performances between centralized GLM and federated GH models. (Left) The centralized GLM with Logistic Regression; (Right) The federated GLMM with 2-degree GH approximation. All methods are applied to all 8 settings of synthetic data that pooled together. Method “centralized GLM”is done with R packages “lm” in a centralized settings
Fig. 4The precision and recall in variable-wise among centralized, Laplace, and Gauss–Hermite method under significance level . (Left) The precision of the test compared to the true value. (Right) The recall of the test compared to the true value
Fig. 6The curve of test power among centralized, Laplace, and Gauss–Hermite methods. (Left) The power of the test of the Laplace method. (Middle) The power of the test of the 2-degree Gauss–Hermite method. (Right) The power of the test of the Centralized method. Power was calculated as the two-sided t-test on p values among different methods
Fig. 5The accuracy in 8 different data settings among centralized, Laplace, and Gauss–Hermite method under significance level . The performance of models in different number of nodes (2 vs 10), sample size (500 vs 30), and variance (large vs small)
The convergence rates on approximation methods LA and GH. (Both LA and GH held the same convergence threshold . The mean values and standard deviations (in parentheses) were given)
| Setting | LA | GH | ||
|---|---|---|---|---|
| Steps | Runtime (s) | Steps | Runtime (s) | |
| 1 | 22.875 (21.623) | 47.953 (20.513) | 34.850 (9.213) | 104.460 (10.614) |
| 2 | 21.500 (21.977) | 40.947 (36.466) | 35.000 (8.711) | 100.940 (19.940) |
| 3 | 29.867 (31.719) | 108.931 (65.486) | 34.900 (6.138) | 1259.285 (231.956) |
| 4 | 27.846 (24.034) | 84.343 (76.502) | 36.650 (6.310) | 1342.695 (250.603) |
| 5 | 59.722 (42.057) | 10.631 (3.945) | 33.750 (10.146) | 12.568 (2.116) |
| 6 | 67.188 (48.994) | 10.499 (4.054) | 31.400 (11.081) | 11.430 (3.064) |
| 7 | 96.286 (53.635) | 96.501 (38.632) | 37.450 (3.818) | 369.165 (41.998) |
| 8 | 116.083 (46.479) | 91.304 (62.410) | 37.150 (4.295) | 309.693 (36.621) |
Statistics of goodness of fit among different methods
| Log-likelihood | AIC | BIC | |
|---|---|---|---|
| R | 13562.9 | 27165.9 | 27340.8 |
| LA | − 13695.0 | 27428.0 | 27594.1 |
| GH | − 11.8 | 61.6 | 227.7 |
Statistics of performances among different methods (95% CIs were generated by Wilson Score interval)
| Precision | Recall | F1-score | AUC | threshold | ||
|---|---|---|---|---|---|---|
| Centralized method with LA | Value | 0.1507 | 0.6204 | 0.2425 | 0.6789 | 0.0900 |
| Lower bound (0.95) | 0.1474 | 0.6160 | 0.2386 | 0.6700 | ||
| Upper bound (0.95) | 0.1539 | 0.6248 | 0.2464 | 0.6878 | ||
| GH with regularization | Value | 0.1705 | 0.6546 | 0.2705 | 0.7178 | 0.0108 |
| Lower bound (0.95) | 0.1670 | 0.6503 | 0.2664 | 0.7091 | ||
| Upper bound (0.95) | 0.1739 | 0.6589 | 0.2745 | 0.7265 |
Fig. 7The ROC curve with Area Under Curve (AUC) among centralized, Laplace, and Gauss–Hermite methods. The orange ROC curve is the centralized method without regularization and the Laplace approximation(i.e., R implementation in the ‘lme4’ package, which does not have an option for including regularization). AUC values are also included, a higher AUC value implicates better performance of the model. The green ROC curve is the 2-degree Gauss–Hermite method with regularization
| Index of sites | Log-likelihood function for site | ||
| Index of patients in a specific site | Parameters of fixed effect | ||
| Index of Hermite polynomial | Parameters of random effect in site | ||
| Order of Hermite polynomial | Hyper-parameters | ||
| Number of sites | Parameter space | ||
| Number of patients in site | A vector represents the data of | ||
| Likelihood function for site | The outcome of patient | ||
| The parameter of regularization term | Number of variables |
| LA | GH |
|---|---|
| Number of variables | Number of variables |
| Scalar | Scalar |
| Scalar | Scalar |
| Scalar first order derivative | Scalar first order derivative |
| Scalar |