| Literature DB >> 33193046 |
Jin Hong1,2, Zhangzhi Feng2, Shui-Hua Wang3,4, Andrew Peet5, Yu-Dong Zhang1,6, Yu Sun5,7, Ming Yang2.
Abstract
Predicting brain age of children accurately and quantitatively can give help in brain development analysis and brain disease diagnosis. Traditional methods to estimate brain age based on 3D magnetic resonance (MR), T1 weighted imaging (T1WI), and diffusion tensor imaging (DTI) need complex preprocessing and extra scanning time, decreasing clinical practice, especially in children. This research aims at proposing an end-to-end AI system based on deep learning to predict the brain age based on routine brain MR imaging. We spent over 5 years enrolling 220 stacked 2D routine clinical brain MR T1-weighted images of healthy children aged 0 to 5 years old and randomly divided those images into training data including 176 subjects and test data including 44 subjects. Data augmentation technology, which includes scaling, image rotation, translation, and gamma correction, was employed to extend the training data. A 10-layer 3D convolutional neural network (CNN) was designed for predicting the brain age of children and it achieved reliable and accurate results on test data with a mean absolute deviation (MAE) of 67.6 days, a root mean squared error (RMSE) of 96.1 days, a mean relative error (MRE) of 8.2%, a correlation coefficient (R) of 0.985, and a coefficient of determination (R 2) of 0.971. Specially, the performance on predicting the age of children under 2 years old with a MAE of 28.9 days, a RMSE of 37.0 days, a MRE of 7.8%, a R of 0.983, and a R 2 of 0.967 is much better than that over 2 with a MAE of 110.0 days, a RMSE of 133.5 days, a MRE of 8.2%, a R of 0.883, and a R 2 of 0.780.Entities:
Keywords: artificial intelligence; brain age; convolutional neural network; deep learning; magnetic resonance imaging
Year: 2020 PMID: 33193046 PMCID: PMC7604456 DOI: 10.3389/fneur.2020.584682
Source DB: PubMed Journal: Front Neurol ISSN: 1664-2295 Impact factor: 4.003
Figure 1Illustration of data augmentation to a 3D image.
Figure 2The hierarchical architecture of proposed 3D CNN. The “32” and “128 × 116 × 12” in “32@128 × 116 × 12” denote the number and size of feature maps.
Figure 3The hierarchical architecture of the 2D CNN.
Figure 4Distribution of participant ages.
Subjects demographic (Std denotes standard deviation).
| Subjects | 88 | 132 | 23 | 21 | 65 | 111 |
| Age (days) | 4–697 | 731–1820 | 36–680 | 749–1687 | 4–697 | 731–1820 |
| Mean | 283.6 ± 215.7 | 1333.5 ± 314.2 | 244.2 ± 201.0 | 1267.0 ± 288.4 | 297.6 ± 220.5 | 1346.1 ± 318.4 |
Figure 5Training performance of one run. (A) Loss against training epoch, and (B) MAE against training epoch. The loss and MAE are the average of all iterations in one epoch.
Figure 6Prediction results of the proposed 3D CNN. The error bar represents the average and standard deviation of the prediction results over 10-run.
Performance of the proposed 3D CNN in predicting children aged.
| 0–2 | 28.9 | 37.0 | 7.8 | 0.983 | 0.967 |
| 2–5 | 110.0 | 133.5 | 8.2 | 0.883 | 0.780 |
| 0–5 | 67.6 | 96.1 | 8.2 | 0.985 | 0.971 |
Figure 7Bland-Altman plots for the proposed 3D CNN. Plot (A–C) denote 0–2, 2–5, 0–5 age groups, respectively.
Figure 8Prediction results of the proposed method without data augmentation. The error bar represents the average and standard deviation of the prediction results over 10-run.
Comparison of the proposed method with and without data augmentation.
| Without data augmentation | 118.3 | 166.7 | 13.7 | 0.963 | 0.926 |
| With data augmentation | 67.6 | 96.1 | 8.2 | 0.985 | 0.971 |
Performance of different network depths.
| 6 | 2 | 75.9 | 102.8 | 9.2 | 0.983 | 0.967 |
| 6 | 3 | 73.8 | 101.6 | 9.0 | 0.984 | 0.968 |
| 6 | 4 | 70.1 | 95.1 | 8.4 | 0.985 | 0.971 |
| 7 | 2 | 80.4 | 110.5 | 10.0 | 0.981 | 0.962 |
| 7 | 3 | 67.6 | 96.1 | 8.2 | 0.985 | 0.971 |
| 7 | 4 | 70.3 | 97.1 | 8.6 | 0.985 | 0.970 |
| 8 | 2 | 73.4 | 106.8 | 9.2 | 0.982 | 0.964 |
| 8 | 3 | 68.7 | 99.5 | 8.2 | 0.984 | 0.969 |
| 8 | 4 | 69.5 | 96.6 | 8.3 | 0.985 | 0.971 |
Figure 9Comparison of different network depths. “6, 2” (“No. of convolution layers, No. of fully connected layers”) denotes 6 convolution layers and 2 fully connected layers.
Comparison of the proposed method with and without batch normalization.
| Without batch normalization | 132.6 | 189.9 | 15.0 | 0.945 | 0.893 |
| With batch normalization | 67.6 | 96.1 | 8.2 | 0.985 | 0.971 |
Figure 10Training performance of one run without batch normalization. (A) Loss against training epoch, and (B) MAE against training epoch. The loss and MAE are the average of all iterations in one epoch.
Comparison of the proposed 3D CNN trained by different batch size and initial learning rate.
| 16 | 0.0000008 | 72.3 | 96.2 | 8.3 | 0.985 | 0.971 |
| 32 | 0.0000008 | 74.1 | 97.4 | 9.2 | 0.985 | 0.971 |
| 64 | 0.0000008 | 67.6 | 96.1 | 8.2 | 0.985 | 0.971 |
| 64 | 0.0000012 | 69.9 | 97.1 | 8.6 | 0.985 | 0.971 |
| 64 | 0.0000004 | 75.0 | 104.8 | 9.3 | 0.983 | 0.966 |
Comparison of 2D CNN and our proposed 3D CNN.
| 2D CNN | 75.1 | 104.6 | 9.2 | 0.982 | 0.965 |
| 3D CNN | 67.6 | 96.1 | 8.2 | 0.985 | 0.971 |
Comparison with state-of-the-art approaches.
| Feature-based developmental model ( | 92 | 8–590 | 72 |
| HRtoF model ( | 50 | 14–797 | 32.1 |
| The proposed 3D CNN | 88 | 4–697 | 28.9 |
| The proposed 3D CNN | 220 | 4–1,820 | 67.6 |