| Literature DB >> 35955125 |
Bocheng Li1, Yunqiu Zhang1, Xusheng Wu2.
Abstract
With the increasingly available electronic health records (EHR), disease prediction has recently gained immense research attention, where an accurate classifier needs to be trained to map the input prediction signals (e.g., symptoms, auxiliary examination results, etc.) to the estimated diseases for each patient. However, most of the current disease prediction models focus on the prediction of a single disease; in the medical field, a patient often suffers from multiple diseases (especially multiple chronic diseases) at the same time. Therefore, multi-disease prediction is of greater significance for patients' early intervention and treatment, but there is no doubt that multi-disease prediction has higher requirements for data extraction ability and greater complexity of classification. In this paper, we propose a novel disease prediction model DLKN-MLC. The model extracts the information in EHR through deep learning combined with a disease knowledge network, quantifies the correlation between diseases through NodeRank, and completes multi-disease prediction. in addition, we distinguished the importance of common disease symptoms, occasional disease symptoms and auxiliary examination results in the process of disease diagnosis. In empirical and comparative experiments on real EHR datasets, the Hamming loss, one-error rate, ranking loss, average precision, and micro-F1 values of the DLKN-MLC model were 0.2624, 0.2136, 0.2190, 88.21%, and 87.86%, respectively, which were better compared with those from previous methods. Extensive experiments on a real-world EHR dataset have demonstrated the state-of-the-art performance of our proposed model.Entities:
Keywords: deep learning; disease prediction; disease prevention; multi label learning
Mesh:
Year: 2022 PMID: 35955125 PMCID: PMC9368602 DOI: 10.3390/ijerph19159771
Source DB: PubMed Journal: Int J Environ Res Public Health ISSN: 1660-4601 Impact factor: 4.614
Experimental environment setting.
| Experimental Environment | Experimental Configuration |
|---|---|
| GPU | GTX 1050TI |
| CPU | E5-2678V3 |
| Development environment | Python3.7.3 TensorFlow1.15.2 |
| Epoch | 20 |
| Optimizer | Adam |
| LSTM learning rate | 0.001 |
| Dropout | 0.5 |
Figure 1Framework of the DLKN-MLC model.
Figure 2Sequence feature extraction framework based on MCBERT-BiLSTM-CRF. Note: “患者上腹疼痛” means “the patient has epigastric pain”.
Figure 3Structure of BERT.
Confusion matrix.
| Confusion Matrix | Predictive Value | ||
|---|---|---|---|
| Positive | Negative | ||
| Actual value | Positive | TP | FN |
| Negative | FP | TN | |
Weight setting group of binary-weighted KN.
| Group | <113> | <123> | <135> | <137> | <139> |
|---|---|---|---|---|---|
| occasional clinical manifestations | 1 | 1 | 1 | 1 | 1 |
| common clinical manifestations | 1 | 2 | 3 | 3 | 3 |
| auxiliary diagnostic results | 3 | 3 | 5 | 7 | 9 |
Experimental results of weight setting of five groups (MEA ± SD).
| HL↓ | OE↓ | RL↓ | AP↑ | Micro-F1↑ | |
|---|---|---|---|---|---|
| <113> | 0.3076 ± 0.005634 | 0.2412 ± 0.008233 | 0.3153 ± 0.009842 | 0.8623 ± 0.005285 | 0.8496 ± 0.003933 |
| <123> | 0.2966 ± 0.005754 | 0.2357 ± 0.009018 | 0.2962 ± 0.009632 | 0.8642 ± 0.005021 | 0.8522 ± 0.003828 |
| <135> | 0.2687 ± 0.004982 | 0.2257 ± 0.008721 | 0.2276 ± 0.010223 | 0.8786 ± 0.004692 | 0.8672 ± 0.003468 |
| <137> |
| 0.2136 ± 0.009728 | 0.2190 ± 0.010373 |
|
|
| <139> | 0.2695 ± 0.005229 |
|
| 0.8754 ± 0.004724 | 0.8654 ± 0.003622 |
Note: The bold value in the table is the optimal value under the index; “↑” means that the larger the index is, the better the classification effect is; “↓” means that the smaller the index is, the better the classification effect is.
Figure 4Model experimental results.
Introduction to comparison model.
| Comparison Model | Model Description |
|---|---|
| Text-CNN | On the basis of CNN, many sliding windows of different sizes are added, and the feature extraction is carried out by a convolution kernel. |
| CNN-RNN | CNN and RNN are combined to extract the local features of the text, and RNN is used to obtain the sequence features and high-order tag correlation of the text. |
| X-BERT | At the same time, tags and input text are used to generate semantic tag clusters to make better use of the dependency relationship between tags for modeling. |
Model performance comparison results (MEA ± SD).
| HL↓ | OE↓ | RL↓ | AP↑ | Micro-F1↑ | |
|---|---|---|---|---|---|
| Text-CNN | 0.3672 ± 0.009621 | 0.3112 ± 0.008635 | 0.2922 ± 0.013585 | 0.7838 ± 0.005145 | 0.7838 ± 0.005785 |
| CNN–RNN | 0.2914 ± 0.006888 | 0.2598 ± 0.009537 | 0.2454 ± 0.008639 | 0.8204 ± 0.005848 | 0.8058 ± 0.007243 |
| X-BERT | 0.2788 ± 0.006675 | 0.2412 ± 0.006431 | 0.2494 ± 0.009972 | 0.8528 ± 0.007514 | 0.8362 ± 0.004946 |
| Our method |
|
|
|
|
|
Note: The bold value in the table is the optimal value under the index; “↑” means that the larger the index is, the better the classification effect is; “↓” means that the smaller the index is, the better the classification effect is.