Literature DB >> 34966294

Multiple Laplacian Regularized RBF Neural Network for Assessing Dry Weight of Patients With End-Stage Renal Disease.

Xiaoyi Guo1,2, Wei Zhou1, Yan Yu1, Yinghua Cai3, Yuan Zhang1, Aiyan Du1, Qun Lu3, Yijie Ding4, Chao Li5.   

Abstract

Dry weight (DW) is an important dialysis index for patients with end-stage renal disease. It can guide clinical hemodialysis. Brain natriuretic peptide, chest computed tomography image, ultrasound, and bioelectrical impedance analysis are key indicators (multisource information) for assessing DW. By these approaches, a trial-and-error method (traditional measurement method) is employed to assess DW. The assessment of clinician is time-consuming. In this study, we developed a method based on artificial intelligence technology to estimate patient DW. Based on the conventional radial basis function neural (RBFN) network, we propose a multiple Laplacian-regularized RBFN (MLapRBFN) model to predict DW of patient. Compared with other model and body composition monitor, our method achieves the lowest value (1.3226) of root mean square error. In Bland-Altman analysis of MLapRBFN, the number of out agreement interval is least (17 samples). MLapRBFN integrates multiple Laplace regularization terms, and employs an efficient iterative algorithm to solve the model. The ratio of out agreement interval is 3.57%, which is lower than 5%. Therefore, our method can be tentatively applied for clinical evaluation of DW in hemodialysis patients.
Copyright © 2021 Guo, Zhou, Yu, Cai, Zhang, Du, Lu, Ding and Li.

Entities:  

Keywords:  RBF networks; dry weight; end-stage renal disease; machine learning; multiple Laplacian regularized model

Year:  2021        PMID: 34966294      PMCID: PMC8711098          DOI: 10.3389/fphys.2021.790086

Source DB:  PubMed          Journal:  Front Physiol        ISSN: 1664-042X            Impact factor:   4.566


Introduction

Dry weight (DW) refers to a patient’s target weight after the end of dialysis (Grassmann et al., 2000; Wabel et al., 2009). After removing excess water from the body, the patient had no facial swelling, wheezing or sitting breathing, edema of both lower limbs, and distended jugular vein (Alexiadis et al., 2016). The patient’s blood pressure, heart rate, breathing, and other vital signs are stable. There are individual differences in the specific value of DW. Good dry weight control can effectively reduce adverse reactions during dialysis. At present, the DW of hemodialysis patients is mainly evaluated by clinical means. This method is labor-intensive and time-consuming, and requires repeated use of various clinical instruments and biological indicators to complete the evaluation. In the past 10 years, a measuring instrument based on human bioelectrical impedance analysis (BIA), called body composition monitor (BCM) (Jiang et al., 2017), has been accurately determining the DW of patients. The above methods require professionals and cannot be processed on a large scale. In recent years, artificial intelligence technology has been widely utilized in the biomedical field (Chen et al., 2019; Cheng et al., 2019; Lin, 2020; Zhang, 2020; Hu et al., 2021a). Artificial neural networks (ANNs) based on back propagation (BP) were employed to evaluate the total water volume of hemodialysis patients. Compared with conventional clinical calculation equations (Chiu et al., 2005), the ANNs obtained better results. Deep learning (Dao et al., 2021; Lv et al., 2021a,b) also made a great contribution to the clinic, including skin cancer (Esteva et al., 2017), breast cancer (Liu J. et al., 2021), and brain diseases (Liu G. et al., 2018; Liu et al., 2019; Bi et al., 2020; Hu et al., 2020, 2021a,b). In biological field, machine learning has been widely used to solve biological problems, including O-GlcNAcylation site prediction (Jia et al., 2018), microbiology analysis (Qu et al., 2019), microRNAs and cancer association prediction (Zeng et al., 2018), lncRNAs (Cheng et al., 2016; Deng et al., 2021), CircRNAs (Fang et al., 2019; Zhao et al., 2019), DNA methylation site (Wei et al., 2018b; Zou et al., 2019; Dai et al., 2020), osteoporosis diagnoses (Su et al., 2020b), function prediction of proteins (Wei et al., 2018a; Wang H. et al., 2019; Deng et al., 2020b; Ding et al., 2020a; Su et al., 2020a), nucleotide binding sites (Ding et al., 2021b), drug complex network analysis (Ding et al., 2019, 2020b,a; Deng et al., 2020a; Han et al., 2021; Liu H. et al., 2021), protein remote homology (Liu B. et al., 2018), electron transport proteins (Ru et al., 2019), and cell-specific replication. In this study, we proposed a novel predictive model based on radial basis function neural network (RBFN). Different from RBFN, multiple Laplacian-regularized RBFN (MLapRBFN) is a multi-view Laplacian regularized model with L2,1-norm, which introduces multiple graph regular items. The Laplacian regular items consider the topological relationship between each patient.

Materials and Methods

Materials

The data set of this study came from the hemodialysis center of Wuxi and the northern Jiangsu People’s Hospital. Our study was approved by the ethics committees (Nos. 2018KY-001 and KYLLKS201813). There are a total of 476 hemodialysis patients. These patients meet the following conditions: over 18 years old; more than 3 months end-stage renal disease (ESRD) and maintenance hemodialysis; diseases such as metal implants, infections, heart failure, disability, pregnancy, and edema, do not appear in the patient population. DW is determined by clinical scoring, which is based on brain natriuretic peptide (BNP), chest computed tomography (CT) image, ultrasound, and bioelectrical impedance analysis (BIA). In addition, age, gender, diastolic blood pressure (DBP), systolic blood pressure (SBP), years of dialysis (YD), heart rate (HR), and body mass index (BMI) are employed to construct our model. The summary of information is shown in Table 1. This study is a retrospective study, and feature of patient is easy to obtain. We hope to provide patients with a non-invasive dry weight assessment method through machine learning models.
TABLE 1

She summary information of patients.

Feature Value r *
Gender (males/females)312/164–0.4489
Age (years)54.17 ± 14.22–0.2341
BMI22.96 ± 2.950.9558
HR (times/min)73.41 ± 8.920.1862
DBP (mmHg)88.32 ± 19.56–0.1249
SBP (mmHg)150.64 ± 29.36–0.1739
YD (years)5.97 ± 3.22–0.1069

*Denotes correlation coefficient between individual variables and dry weight value.

She summary information of patients. *Denotes correlation coefficient between individual variables and dry weight value.

Radial Basis Function Network

The RBF neural network (Park and Sandberg, 1990) is composed of an input layer, a hidden layer, and an output layer, and is shown in Figure 1. The feature of the hemodialysis patient can be fed into the RBF network for processing. Suppose that there are N samples containing d variables ({x, y}, i = 1, 2, …, N). The output is y ∈ R1× and input data x ∈ R1×. In Figure 1, the RBF network has three layers, which are the input, hidden, and output layers. The transformation from the input space to the hidden layer space is non-linear, and the transformation between the hidden layer and the output layer is linear. The fundamental of the RBFN is the RBF is the “base” of the hidden unit. The vector of input layer can be mapped to the space of the hidden layer without weight connection. The mapping relationship is determined with center point of hidden unit. In Figure 1, the number of hidden layer nodes is p (p center points). The activation function of the RBF neural network can be represented as:
FIGURE 1

Topological structure of the radial basis function (RBF) network.

where x is a feature vector of sample, c, j ∈ {1, 2, …, p} is the vector of j-th center point, and σ is the width parameter of the function, which controls the radial range of the function. RBFN can be represented as: where matrix Φ ∈ R is the output of the hidden nodes. W ∈ R is the weight matrix between output and hidden nodes. Y ∈ R is the matrix of dependent variable. Φ ∈ R and W ∈ R can be represented as: Topological structure of the radial basis function (RBF) network. To train RBFN, 3 parameters should be solved: the center points of the basis function (c ∈ R1×, j ∈ {1, 2, …, p}), width (σ), and weight matrix between the output and hidden layer (W ∈ R). In most cases, the self-organized selection center learning method is used: (1) unsupervised learning process, solving the center points and width of RBF, (2) supervised learning process, solving the weights W ∈ R. First, select p centers for k-means clustering (p clusters). For the radial basis of the Gaussian kernel function, the width is solved by the formula: where σmax is the maximum distance between the selected center points. For W ∈ R, the RBF network can be represented as: The gradient of Eq. 6 can be set as 0: For a new test sample x, we can estimate as: where Φ = (ϕ (x, c1), ϕ (x, c2), …, ϕ (x, c))1 × .

Proposed Model of Multiple Laplacian-Regularized RBF Network

To further improve the generalization performance of the RBF network, a multi-view Laplacian-regularized RBF network is proposed. The unsupervised process of the first stage remains unchanged; we mainly improve the model in the second part. The objective function of Eq. 6 is revised as: where λ1 and λ2 denote the parameters of Laplacian and L2, 1-norm term. L2, 1-norm can be used to obtain a sparser solution during the training process, which makes the model more robust. ρ > 1, which is used to prevent the extreme situations of η = 0 (or η = 1). L ∈ R is the Laplacian matrix, which is employed to represent the manifold of samples. V is the number of views. L is built by heat kernel matrix S ∈ R: In our study, we employ the following functions to construct the heat kernel matrix: where γ is a constant, and we set it as 1.

Optimization

The third term of cannot be diversified, so Eq. 9 is converted into: where G ∈ R is a diagonal matrix, and i-th is an element: Since there are multiple variables (W, η) that need to be optimized, we first fix η and optimize W. We initialize , and get the fused matrix . Equation 12 can be written as: We obtain the derivative of formula (14) for variable W: Then, we fix the variant W and optimize η, v = 1, 2, …, V, which is related to: The above formula can be converted to a Lagrange function: We set η and ξ to 0: where η can be estimated by: The process of MLapRBFN is listed in Algorithm 1 and Figure 2.
FIGURE 2

Flow chart of multiple Laplacian-regularized RBF network (MLapRBFN).

Require: Training set {x, y}, i = 1, 2, …, N, new samples{x}, j = 1, 2, …, M, the hidden layer nodes (p), the iterations tmax, parameters of λ1and λ2; Ensure: The predictive values of (1) Using k-means to select p centers and width (σ ) for RBF function. Calculating the W (training set) and Laplacian matrices L, v = 1, 2, …, V by Eqs 1, 10, and 11. Initializing ; (2) Setting t = 0, estimate the initial W(0) with Eq. 7; Repeat (3) Update the G(with (4) Update W( via Eq. 15d; (5) Update via Eq. 19; (6) Calculating until t > tmax; (7) Calculate the output matrix Φ(test set); (8) Predict {y}, j = 1, 2, …, M by . Flow chart of multiple Laplacian-regularized RBF network (MLapRBFN).

Results and Discussion

In this section, we perform 10-fold cross-validation (10-CV) to test the predictive performance of different models, the BCM device, multiple kernel support vector regression (MKSVR), linear regression (LR), back propagation-based artificial neural network [ANN (BP)], and multi-kernel ridge regression (MKRR).

Evaluation Measurements

Some traditional assessment methods include correlation coefficient (R), R square, root mean square error (RMSE), empirical cumulative distribution plot (ECDP), and Bland–Altman analysis. In particular, The Bland–Altman analysis usually can evaluate the agreement between two methods, and determines whether the two methods can be replaced with each other.

Selection of Optimal Parameters

We obtain the optimal model parameters, such as the number of hidden layer nodes (the number of clusters), regularization parameters, and number of optimization iterations. First, we fix the number of iterations (tmax = 10) and regularization coefficients (λ1 = 20, λ2 = 20), and adjust the number of hidden layer neurons. In Figure 3A, we test the number of nodes from 10 to 250 with step of 10. After adding 140 neuron nodes, RMSE tends to be flat. In addition, the RMSE also tends to remain unchanged (minimum) after 10 iterations (Figure 3B). Finally, we get the number of hidden layer nodes as 140 and times of iteration as 10, and adjust the regularization coefficient. The results are shown in Figure 4, and the optimal coefficients are .
FIGURE 3

RMSEs under different parameters. (A) Root mean square errors (RMSEs) under different hidden layer nodes. (B) RMSEs under different numbers of iteration.

FIGURE 4

Root mean square errors under different regularization coefficients.

RMSEs under different parameters. (A) Root mean square errors (RMSEs) under different hidden layer nodes. (B) RMSEs under different numbers of iteration. Root mean square errors under different regularization coefficients.

Comparison With Other Existing Models

We compare our model with other existing machine learning methods (Guo et al., 2021), which include multi-kernel ridge regression (MKRR), multiple kernel support vector regression (MKSVR), artificial neural network (ANN), linear regression (LR), and BCM measuring instrument. The gold standard is clinical dry weight. In Table 2, our method (MLapRBFN) is better than the ordinary RBF neural network (RBFN) model. The R and R squared have the highest values of 0.9511 and 0.9432, respectively. In addition, the RMSE of MLapRBFN reaches its lowest value (1.3226). In Table 2 and Figure 5, we can see that our method has the smallest ECDP range (from −3.8174 to 3.4383). The multiple Laplacian regularized RBFN model has multiple graphs for different heat kernels, which contain different information. To effectively integrate each graph of feature into one graph, multi-view learning (MVL) is employed to estimate the weight of each graph. Each graph has different contribution for the model.
TABLE 2

Comparison with other methods (10-CV).

Method R R SquaredRMSEEmpirical cumulative distribution plot
Highest valueLowest valueMedian value
MKSVR*0.94120.93211.38174.3962−4.1273 0.0082
MKRR*0.93990.92891.50154.9227−4.26040.1104
ANN (BP)*0.93980.92951.47947.3661−4.74470.1324
LR*0.94030.93081.43354.2524−.40140.1418
BCM*0.94730.91371.96943.2235−6.2776−.9863
RBFN0.94100.93021.45144.9018−3.93760.0966
MLapRBFN (our method) 0.9511 0.9432 1.3226 3.4383-3.81740.0822

*The results are from previous work of MKSVR. Bold values represents the best performance for each column.

FIGURE 5

Folded empirical cumulative distribution curves of six methods.

Comparison with other methods (10-CV). *The results are from previous work of MKSVR. Bold values represents the best performance for each column. Folded empirical cumulative distribution curves of six methods.

Bland-Altman Analysis

In this section, we employ a Bland–Altman plot to further evaluate the regression error of different models. In Figure 6 and Table 3, several models evaluate the agreement with clinical DW. MLapRBFN achieves the smallest range of -0.2413 to 0.1601 (95% confidence interval). What is more, the number (ratio) of out agreement interval is the key indicator to evaluate whether the two methods are equivalent. For the predictive models, the number (ratio) should be all less than 24 (5%). In Table 3, all the predictive models meet this standard. In particular, our method obtains the least number (ratio) of 17 (3.57%). If 95% of the samples are in agreement range, the predictive models are clinically acceptable. It can be seen that our method can replace clinical methods.
FIGURE 6

Bland–Altman plot analysis. (A) ANN, (B) LR, (C) MKRR, (D) MKSVR, (E) BCM, and (F) MLapRBFN.

TABLE 3

Bland–Altman plot analysis of the models.

ModelDifferences with DW (%)
Limits of agreement (%)
MeanSD95% confidence intervalLower limitUpper limitNumber (ratio) of out agreement interval
MKSVR*−0.26382.3372−0.4743 to -0.05329−4.84464.317122/476 (4.62%)
MKRR*−0.08012.5007−0.3053 to 0.1451−4.98144.821223/476 (4.83%)
ANN (BP)*0.11522.5139−0.1112 to 0.3416−4.81195.042422/476 (4.62%)
LR*0.00022.4269−0.2184 to 0.2187−4.75664.756921/476 (4.41%)
BCM*−1.82322.7466−2.0706 to −1.5759−7.20663.560130/476 (6.30%)
MLapRBFN (our method)−0.040612.2280−0.2413 to 0.1601−4.40754.326317/476 (3.57%)

*The results are from previous work of MKSVR (

Bland–Altman plot analysis. (A) ANN, (B) LR, (C) MKRR, (D) MKSVR, (E) BCM, and (F) MLapRBFN. Bland–Altman plot analysis of the models. *The results are from previous work of MKSVR (

Conclusion

The limitations of BCM and clinical are time-consuming and laborious. In our study, a MLapRBFN method is developed to predict the DW of hemodialysis patients. Different from standard RBFN, our method contains multiple Laplace regularization terms, and uses an efficient iterative algorithm to solve the model. MKRR, LR, MKSVR, and ANN are compared with our model. Bland-Altman analysis and RMSE are the main evaluation methods. In the Bland-Altman analysis of MLapRBFN, the number of out agreement interval is the least (17 samples). In the fields of medicine (Esteva et al., 2017; Xiao et al., 2017; Hu et al., 2018; Deng et al., 2019; Huang et al., 2020; Zhou et al., 2020; Yang H. et al., 2021), pharmacy (Wang et al., 2020), and biology (Wei et al., 2014, 2017, 2019; Fajila, 2019; Wang Y. et al., 2019; Wang et al., 2021; Yang C. et al., 2021; Zou et al., 2021), artificial intelligence technology has solved lots of predictive tasks. In future studies, more data of hemodialysis patients will be collected, and a deep neural network (Dao et al., 2020a,b; Lv et al., 2020) with stronger representation ability to accurately estimate the DW of hemodialysis patients will be built.

Data Availability Statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding authors.

Ethics Statement

The studies involving human participants were reviewed and approved by the experimental protocol was established and approved by the Human Ethics Committee (Wuxi and Northern Jiangsu People’s Hospital Ethics Committee). The ethical approval numbers are 2018KY-001 and KYLLKS201813. The patients/participants provided their written informed consent to participate in this study.

Author Contributions

XG and WZ did the experiments and wrote the manuscript. YD, CL, and XG designed the method. YY, YZ, AD, QL, and YC revised the manuscript. All authors have read and approved the final manuscript.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
  46 in total

1.  SubtypeDrug: a software package for prioritization of candidate cancer subtype-specific drugs.

Authors:  Xudong Han; Qingfei Kong; Chonghui Liu; Liang Cheng; Junwei Han
Journal:  Bioinformatics       Date:  2021-01-18       Impact factor: 6.937

2.  Pathway-Guided Deep Neural Network toward Interpretable and Predictive Modeling of Drug Sensitivity.

Authors:  Lei Deng; Yideng Cai; Wenhao Zhang; Wenyi Yang; Bo Gao; Hui Liu
Journal:  J Chem Inf Model       Date:  2020-08-31       Impact factor: 4.956

3.  A computational platform to identify origins of replication sites in eukaryotes.

Authors:  Fu-Ying Dao; Hao Lv; Hasan Zulfiqar; Hui Yang; Wei Su; Hui Gao; Hui Ding; Hao Lin
Journal:  Brief Bioinform       Date:  2021-03-22       Impact factor: 11.622

4.  Dermatologist-level classification of skin cancer with deep neural networks.

Authors:  Andre Esteva; Brett Kuprel; Roberto A Novoa; Justin Ko; Susan M Swetter; Helen M Blau; Sebastian Thrun
Journal:  Nature       Date:  2017-01-25       Impact factor: 49.962

5.  Gene Therapy for Hemophilia A: Where We Stand.

Authors:  Miaojin Zhou; Zhiqing Hu; Chunhua Zhang; Lingqian Wu; Zhuo Li; Desheng Liang
Journal:  Curr Gene Ther       Date:  2020       Impact factor: 4.391

6.  Use of lung ultrasonography to determine the accuracy of clinically estimated dry weight in chronic hemodialysis patients.

Authors:  Chuan Jiang; Satyam Patel; Andrew Moses; Maria V DeVita; Michael F Michelis
Journal:  Int Urol Nephrol       Date:  2017-10-03       Impact factor: 2.370

7.  Identification of Alzheimer's Disease-Related Genes Based on Data Integration Method.

Authors:  Yang Hu; Tianyi Zhao; Tianyi Zang; Ying Zhang; Liang Cheng
Journal:  Front Genet       Date:  2019-01-25       Impact factor: 4.599

8.  Melatonin inhibiting the survival of human gastric cancer cells under ER stress involving autophagy and Ras-Raf-MAPK signalling.

Authors:  Yongye Huang; Kexun Yuan; Meifang Tang; Jiaming Yue; Lijun Bao; Shuang Wu; Yanxin Zhang; Yin Li; Yihang Wang; Xu Ou; Jiaxin Gou; Qi Zhao; Lin Yuan
Journal:  J Cell Mol Med       Date:  2020-12-25       Impact factor: 5.310

9.  Integration of Multiple-Omics Data to Analyze the Population-Specific Differences for Coronary Artery Disease.

Authors:  Yang Hu; Shizheng Qiu; Liang Cheng
Journal:  Comput Math Methods Med       Date:  2021-08-17       Impact factor: 2.238

Review 10.  Application of Machine Learning in Microbiology.

Authors:  Kaiyang Qu; Fei Guo; Xiangrong Liu; Yuan Lin; Quan Zou
Journal:  Front Microbiol       Date:  2019-04-18       Impact factor: 5.640

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.