| Literature DB >> 33198170 |
Dae-Yeon Kim1, Dong-Sik Choi2, Jaeyun Kim3, Sung Wan Chun1, Hyo-Wook Gil1, Nam-Jun Cho1, Ah Reum Kang4, Jiyoung Woo3.
Abstract
In this study, we propose a personalized glucose prediction model using deep learning for hospitalized patients who experience Type-2 diabetes. We aim for our model to assist the medical personnel who check the blood glucose and control the amount of insulin doses. Herein, we employed a deep learning algorithm, especially a recurrent neural network (RNN), that consists of a sequence processing layer and a classification layer for the glucose prediction. We tested a simple RNN, gated recurrent unit (GRU), and long-short term memory (LSTM) and varied the architectures to determine the one with the best performance. For that, we collected data for a week using a continuous glucose monitoring device. Type-2 inpatients are usually experiencing bad health conditions and have a high variability of glucose level. However, there are few studies on the Type-2 glucose prediction model while many studies performed on Type-1 glucose prediction. This work has a contribution in that the proposed model exhibits a comparative performance to previous works on Type-1 patients. For 20 in-hospital patients, we achieved an average root mean squared error (RMSE) of 21.5 and an Mean absolute percentage error (MAPE) of 11.1%. The GRU with a single RNN layer and two dense layers was found to be sufficient to predict the glucose level. Moreover, to build a personalized model, at most, 50% of data are required for training.Entities:
Keywords: continuous glucose monitoring; deep learning; diabetic inpatient; glucose prediction model
Mesh:
Substances:
Year: 2020 PMID: 33198170 PMCID: PMC7696446 DOI: 10.3390/s20226460
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Research framework.
Figure 2Conceptual description of a recurrent neural network (RNN) (Simple RNN on the left and gated recurrent unit (GRU) on the right).
Figure 3Conceptual description of long-short term memory (LSTM).
Figure 4Overall description of how the data are framed to be fitted to the RNN model.
Figure 5Form of the input sample.
Demographics of enrolled patients.
| Age Group | Gender | ||
|---|---|---|---|
| 30–39 | 3 | Female | 13 |
| 40–49 | 6 | ||
| 50–59 | 4 | Male | 7 |
| 60–69 | 7 | ||
Distribution of the glucose levels of patients.
| Patient ID | Experiment Days | Max | Min | Average |
|---|---|---|---|---|
| S001 | 5.9 | 350 | 73 | 186.3 |
| S002 | 3.6 | 400 | 78 | 167.61 |
| S003 | 4.4 | 400 | 60 | 200.73 |
| S004 | 6.7 | 334 | 47 | 155.02 |
| S005 | 6.9 | 347 | 40 | 164.45 |
| S006 | 7.4 | 400 | 117 | 229.4 |
| S007 | 5.8 | 301 | 71 | 153.78 |
| S008 | 4.5 | 398 | 60 | 146.41 |
| S009 | 6.3 | 506 | 93 | 217.22 |
| S010 | 2.6 | 400 | 169 | 278.03 |
| S011 | 3.6 | 400 | 62 | 224.44 |
| S012 | 4.9 | 372 | 56 | 196.72 |
| S013 | 2.8 | 324 | 60 | 193.54 |
| S014 | 3.9 | 288 | 67 | 154.98 |
| S015 | 6.2 | 403 | 88 | 212.51 |
| S016 | 3.7 | 289 | 80 | 172.45 |
| S017 | 6.8 | 399 | 62 | 204.75 |
| S018 | 4.3 | 399 | 145 | 248.36 |
| S019 | 3.1 | 275 | 79 | 131.22 |
| S020 | 3.9 | 400 | 62 | 215.09 |
| Total | 99.4 | 400 | 60 | 193.1 |
Algorithm comparison.
| RNN | GRU | LSTM | |
|---|---|---|---|
| RMSE 7:3 with batch size 20 | 23.43873 | 22.26309 | 23.08421 |
Model architecture comparison.
| Models | RMSE |
|---|---|
| One layer in sequence processing | 22.26309 |
| Bi-directional | 22.41404 |
| Two layers | 22.5565 |
Training sample size comparison.
| Train: Test | GRU Performance |
|---|---|
| 7:3 | 22.26309 |
| 6:4 | 21.79546 |
| 5:5 | 22.11716 |
| 4:6 | 22.29638 |
| 3:7 | 22.57999 |
Model improvement results.
| Hyper-Parameters | Performance |
|---|---|
| Batch 50 | 22.37776 |
| Batch 50 + Shuffling | 21.46406 |
| Batch 50 + With Shuffle + Adamax optimization | 21.47056 |
| Batch 50 + With Shuffle + RMSprop optimization + 10 epoch | 22.45610 |
| Batch 50 + With Shuffle + RMSprop optimization + 30 epoch | 24.21941 |
Figure 6The worst case in the first panel and the best case in the second panel. The dotted line is the start time of testing. The red line is for the reference value and the green line is for the prediction value.
Figure 7Error analysis with two axes—actual value and predicted value.
Percentage of data points in grid zones.
| Zone | A | B | C | D | E |
|---|---|---|---|---|---|
| Percentage | 87.92 | 11.11 | 0.06 | 0.91 | 0 |