| Literature DB >> 31068968 |
Xiaoqing Zhang1, Hongling Zhao1, Shuo Zhang1, Runzhi Li1.
Abstract
Chronic diseases are one of the biggest threats to human life. It is clinically significant to predict the chronic disease prior to diagnosis time and take effective therapy as early as possible. In this work, we use problem transform methods to convert the chronic diseases prediction into a multi-label classification problem and propose a novel convolutional neural network (CNN) architecture named GroupNet to solve the multi-label chronic disease classification problem. Binary Relevance (BR) and Label Powerset (LP) methods are adopted to transform multiple chronic disease labels. We present the correlated loss as the loss function used in the GroupNet, which integrates the correlation coefficient between different diseases. The experiments are conducted on the physical examination datasets collected from a local medical center. In the experiments, we compare GroupNet with other methods and models. GroupNet outperforms others and achieves the best accuracy of 81.13%.Entities:
Keywords: GroupNet; chronic disease; correlated loss; group block; multi-label classification
Year: 2019 PMID: 31068968 PMCID: PMC6491565 DOI: 10.3389/fgene.2019.00351
Source DB: PubMed Journal: Front Genet ISSN: 1664-8021 Impact factor: 4.599
Figure 1(A) Distribution of multiple chronic diseases; (B) Distribution of single-label of three chronic diseases dependencies; (C) Correlation coefficient matrix of three types of chronic diseases (hypertension, diabetes, and fatty liver), and they are computed by Pearson product-moment correlation coefficient.
Figure 2Group convolution strategy.
Figure 3The paradigm of Group Block. For L continuous convolution layers, M and N denotes the number of independent partition convolution units.
Figure 4GroupNet Architecture.
Figure 5(A) LP-GroupNet, (B) LP-GroupNet-3, (C) LP-GroupNet-4.
Figure 6(A) relationship between accuracy and epochs; (B) relationship between accuracy and learning rate; (C) relationship between accuracy and batch size.
Figure 7(A) Relationship between performance and convolution kernel size; (B) Relationship between performance and dropout rate; (C) Relationship between performance and activation function; (D) Relationship between performance and focusing parameter γ in focal loss. Blue denotes accuracy and red denotesF1.
Comparison of Adam and SGD.
| SGD | 75.09 | 74.50 | 75.09 | 74.50 |
| Adam | 79.77 | 79.84 | 79.77 | 79.40 |
Comparison of different number of partition convolution units in group block.
| LP-GroupNet | 79.77 | 79.84 | 79.77 | 79.40 |
| GroupNet-3 | 79.66 | 79.42 | 79.66 | 79.22 |
| GroupNet-4 | 79.20 | 78.88 | 78.20 | 78.88 |
Hyper-parameter settings of the GroupNet.
| Learning rate | 0.002 |
| Epochs | 20 |
| Batch size | 128 |
| Convolution kernel size | 1 × 3 |
| Dropout rate | 0.5 |
| Activation Function | tanh |
| γ | 2 |
| Optimizer | Adam |
| The number of partition convolution units | 2 |
Comparison of CNN models based on LP method.
| GroupNet | 79.77 | 79.84 | 79.77 | 79.40 |
| IGCNet | 78.08 | 77.64 | 78.08 | 77.65 |
| GoogleNet | 78.56 | 79.02 | 78.56 | 78.41 |
| AlexNet | 76.28 | 77.03 | 76.28 | 76.10 |
| VGGNet | 78.17 | 77.79 | 78.17 | 77.46 |
Comparison of LP-GroupNet and BR-GroupNet.
| LP-GroupNet | 79.77 | 79.84 | 79.77 | 79.40 |
| BR-GroupNet | 80.54 | 80.70 | 80.54 | 80.35 |
Comparison of different loss functions based on the BR-GroupNet.
| CE | 79.05 | 78.77 | 79.05 | 78.54 |
| FL | 80.54 | 80.70 | 80.54 | 80.35 |
| CL1 | 79.66 | 80.59 | 79.66 | 79.30 |
| CL2 | 81.13 | 81.37 | 81.13 | 81.02 |
Comparison of GroupNet model and other comparative methods.
| BR-GroupNet-CL | 81.13 | 81.37 | 81.13 | 81.02 |
| IGCNet | 78.08 | 77.64 | 78.08 | 77.65 |
| GoogleNet | 78.56 | 79.02 | 78.56 | 78.41 |
| AlexNet | 76.28 | 77.03 | 76.28 | 76.10 |
| VGGNet | 78.17 | 77.79 | 78.17 | 77.46 |
| DNN | 71.10 | 75.70 | 71.12 | 72.61 |
| LSTM | 75.83 | 75.31 | 75.83 | 75.24 |
| GRU | 76.35 | 76.34 | 76.35 | 75.58 |
| DT | 77.26 | 77.12 | 77.34 | 77.12 |
| MLP | 74.94 | 74.40 | 74.95 | 74.40 |
| SVM | 48.89 | 42.2 | 49.91 | 41.6 |
| SMO | 70.12 | 67.60 | 70.12 | 67.42 |
| ML-KNN | 51.03 | 60.21 | 53.02 | 50.47 |
| BPMLL | 76.65 | 76.72 | 76.65 | 76.32 |