| Literature DB >> 31031657 |
Shui-Hua Wang1,2,3, Shipeng Xie4, Xianqing Chen5, David S Guttery3, Chaosheng Tang1, Junding Sun1, Yu-Dong Zhang1,3,6.
Abstract
Aim: This paper proposes a novel alcoholism identification approach that can assist radiologists in patient diagnosis. Method: AlexNet was used as the basic transfer learning model. The global learning rate was small, at 10-4, and the iteration epoch number was at 10. The learning rate factor of replaced layers was 10 times larger than that of the transferred layers. We tested five different replacement configurations of transfer learning.Entities:
Keywords: AlexNet; alcoholism; convolutional neural network; data augmentation; dropout; local response normalization; magnetic resonance imaging; transfer learning
Year: 2019 PMID: 31031657 PMCID: PMC6470295 DOI: 10.3389/fpsyt.2019.00205
Source DB: PubMed Journal: Front Psychiatry ISSN: 1664-0640 Impact factor: 4.157
Dataset division into training, validation, and test sets.
| Training | 80 | 80 | 160 |
| Validation | 30 | 30 | 60 |
| Test | 78 | 81 | 159 |
| Total | 188 | 191 | 379 |
Data augmentation.
| Original Image | 80 | 80 | 160 |
| DA_I: Noise Injection | 2,400 | 2,400 | 4,800 |
| DA_II: Scaling | 2,400 | 2,400 | 4,800 |
| DA_III: Random Translation | 2,400 | 2,400 | 4,800 |
| DA_IV: Image Rotation | 2,400 | 2,400 | 4,800 |
| DA_V: Gamma Correction | 2,400 | 2,400 | 4,800 |
| Horizontal-flipped Image | 80 | 80 | 160 |
| DA_I: Noise Injection | 2,400 | 2,400 | 4,800 |
| DA_II: Scaling | 2,400 | 2,400 | 4,800 |
| DA_III: Random Translation | 2,400 | 2,400 | 4,800 |
| DA_IV: Image Rotation | 2,400 | 2,400 | 4,800 |
| DA_V: Gamma Correction | 2,400 | 2,400 | 4,800 |
| New Training Data | 24,160 | 24,160 | 48,320 |
Figure 1Idea of transfer learning.
Figure 2Structure of AlexNet (5 CLs and 3 FCLs).
Learnable layers in AlexNet.
| CL1 | 11*11*3*96 = 34,848 | 1*1*96 = 96 |
| CL2 | 5*5*48*256 = 307,200 | 1*1*256 = 256 |
| CL3 | 3*3*256*384 = 884,736 | 1*1*384 = 384 |
| CL4 | 3*3*192*384 = 663,552 | 1*1*384 = 384 |
| CL5 | 3*3*192*256 = 442,368 | 1*1*256 = 256 |
| FCL6 | 4096*9216 = 37,748,736 | 4096*1 = 4,096 |
| FCL7 | 4096*4096 = 16,777,216 | 4096*1 = 4,096 |
| FCL8 | 1000*4096 = 4,096,000 | 1000*1 = 1,000 |
| CL Subtotal | 2,332,704 | 1,376 |
| FCL Subtotal | 58,621,952 | 9,192 |
| Total | 60,954,656 | 10,568 |
Figure 3Illustration of convolution operation.
Figure 4Example of max pooling (stride = 2, kernels size = 2).
Figure 5Structure of last fully-connected layer (C stands for the number of total classes).
Figure 6Two modes of activation function. (A) Single input single output mode. (B) Multiple input multiple output mode.
Figure 7Dropout neural network. (A) Before dropout. (B) After dropout.
Revision of Last three layers of AlexNet.
| 23 | FCL (1000) with pre-trained weights and biases | FCL (2) with random initialization |
| 24 | Softmax Layer | Softmax Layer |
| 25 | Classification Layer | Classification Layer |
Figure 8Five different settings A-E (Setting A stands for the layers from first layer till layer A are transferred layers, and the remaining layers are replaced layers).
Pseudocode of our experiment.
| [NonTest, Test] = split(Dataset); |
| for S = [A, B, C, D, E] |
| for i = 1:10 |
| [train(i), valid(i)] = split(NonTest), |
| Model(S, i) = TrainNetwork(AlexNet, train(i), valid(i), Setting = S), |
| PerfValid(S, i) = Predict(Model(S, i), valid(i)), |
| end |
| PerfValid(S) = mean(PerfValid(S, i)), |
| End |
| S* = argmax[Performance(S)], |
| for i = 1:10 |
| [train(i), valid(i)] = split(NonTest), |
| Model(S*, i) = TrainNetwork(AlexNet, train(i), valid(i), Setting = S*), |
| PerfTest(S*, i) = predict(Model(S*, i), Test), |
| End |
| PerfTest(S*) = mean(PerfTest(S*, i)), |
| Output PerfTest(S*), |
Figure 9Data augmentation by horizontal flipping. (A) Original image. (B) Flipped image.
Figure 10Five augmentation techniques of the original image. (A) Noise injection. (B) Scaling. (C) Random translation. (D) Image rotation. (E) Gamma correction.
Ten runs of validation performance of transfer learning using Setting A.
| 1 | 96.67 | 93.33 | 93.54 | 95.00 | 95.05 |
| 2 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
| 3 | 90.00 | 100.00 | 100.00 | 95.00 | 94.70 |
| 4 | 96.67 | 90.00 | 90.63 | 93.33 | 93.55 |
| 5 | 90.00 | 96.67 | 96.43 | 93.33 | 93.10 |
| 6 | 96.67 | 96.67 | 96.67 | 96.67 | 96.67 |
| 7 | 96.67 | 96.67 | 96.88 | 96.67 | 96.66 |
| 8 | 96.67 | 100.00 | 100.00 | 98.33 | 98.28 |
| 9 | 100.00 | 90.00 | 90.99 | 95.00 | 95.26 |
| 10 | 96.67 | 93.33 | 93.54 | 95.00 | 95.05 |
| Mean ± | 96.00 ± 3.27 | 95.67 ± 3.67 | 95.87 ± 3.40 | 95.83 ± 2.01 | 95.83 ± 2.00 |
Ten runs of validation performance of transfer learning using Setting E.
| 1 | 93.33 | 100.00 | 100.00 | 96.67 | 96.55 |
| 2 | 100.00 | 96.67 | 96.88 | 98.33 | 98.39 |
| 3 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
| 4 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
| 5 | 93.33 | 93.33 | 93.33 | 93.33 | 93.33 |
| 6 | 96.67 | 100.00 | 100.00 | 98.33 | 98.28 |
| 7 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
| 8 | 96.67 | 93.33 | 93.54 | 95.00 | 95.05 |
| 9 | 96.67 | 93.33 | 93.54 | 95.00 | 95.05 |
| 10 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
| Mean ± | 97.67 ± 2.60 | 97.67 ± 3.00 | 97.73 ± 2.93 | 97.67 ± 2.38 | 97.67 ± 2.37 |
Comparison of different setting.
| A | 96.00 ± 3.27 | 95.67 ± 3.67 | 95.87 ± 3.40 | 95.83 ± 2.01 | 95.83 ± 2.00 |
| B | 96.33 ± 3.79 | 96.00 ± 2.49 | 96.12 ± 2.43 | 96.17 ± 2.36 | 96.15 ± 2.43 |
| C | 96.33 ± 3.48 | 96.33 ± 3.14 | 96.49 ± 2.94 | 96.33 ± 2.08 | 96.33 ± 2.11 |
| D | 97.00 ± 3.79 | 97.00 ± 2.77 | 97.06 ± 2.70 | 97.00 ± 2.56 | 96.98 ± 2.62 |
Bold means the best.
Figure 11Error bar of five TL settings.
Learnable layers in optimal transfer learning model.
| CL1 (Ours) | 11*11*3*96 = 34,848 | 0.06 | 1*1*96 = 96 | 1.00 |
| CL2 (Ours) | 5*5*48*256 = 307,200 | 0.54 | 1*1*256 = 256 | 2.68 |
| CL3 (Ours) | 3*3*256*384 = 884,736 | 1.56 | 1*1*384 = 384 | 4.01 |
| CL4 (Ours) | 3*3*192*384 = 663,552 | 1.17 | 1*1*384 = 384 | 4.01 |
| CL5 (Ours) | 3*3*192*256 = 442,368 | 0.78 | 1*1*256 = 256 | 2.68 |
| FCL6 (Ours) | 4096*9216 = 37,748,736 | 66.38 | 4096*1 = 4,096 | 42.80 |
| FCL7 (Ours) | 4096*4096 = 16,777,216 | 29.50 | 4096*1 = 4,096 | 42.80 |
| FCL8 (AlexNet) | 1000*4096 = 4,096,000 | 1000*1 = 1,000 | ||
| FCL8 (Ours) | 2*4096 = 8,192 | 0.01 | 2*1 = 2 | 0.02 |
| CL Subtotal (AlexNet) | 2,332,704 | 1,376 | ||
| CL Subtotal (Ours) | 2,332,704 | 4.10 | 1,376 | 14.38 |
| FCL Subtotal (AlexNet) | 58,621,952 | 9,192 | ||
| FCL Subtotal (Ours) | 54,534,144 | 95.90 | 8,194 | 85.62 |
| Total (AlexNet) | 60,954,656 | 10,568 | ||
| Total (Ours) | 56,866,848 | 100 | 9,570 | 100 |
Ten runs without using data augmentation (Setting E).
| 1 | 83.33 | 96.67 | 96.15 | 90.00 | 89.29 |
| 2 | 96.67 | 93.33 | 93.54 | 95.00 | 95.05 |
| 3 | 96.67 | 93.33 | 93.54 | 95.00 | 95.05 |
| 4 | 96.67 | 90.00 | 90.78 | 93.33 | 93.54 |
| 5 | 96.67 | 100.00 | 100.00 | 98.33 | 98.28 |
| 6 | 96.67 | 96.67 | 96.67 | 96.67 | 96.67 |
| 7 | 96.67 | 93.33 | 93.54 | 95.00 | 95.05 |
| 8 | 93.33 | 100.00 | 100.00 | 96.67 | 96.55 |
| 9 | 93.33 | 96.67 | 96.67 | 95.00 | 94.94 |
| 10 | 100.00 | 93.33 | 93.75 | 96.67 | 96.77 |
| Mean ± | 95.00 ± 4.28 | 95.33 ± 3.06 | 95.46 ± 2.84 | 95.17 ± 2.17 | 95.12 ± 2.32 |
Effect of using data augmentation technique.
| Not use DA | 95.00 ± 4.28 | 95.33 ± 3.06 | 95.46 ± 2.84 | 95.17 ± 2.17 | 95.12 ± 2.32 |
| Use DA (ours) | 97.67 ± 2.60 | 97.67 ± 3.00 | 97.73 ± 2.93 | 97.67 ± 2.38 | 97.67 ± 2.37 |
Ten runs of proposed method on the test set (Setting E).
| 1 | 97.44 | 96.31 | 96.22 | 96.86 | 96.82 |
| 2 | 98.72 | 93.81 | 93.93 | 96.23 | 96.25 |
| 3 | 94.87 | 96.31 | 96.09 | 95.61 | 95.47 |
| 4 | 97.44 | 98.75 | 98.72 | 98.11 | 98.07 |
| 5 | 98.72 | 98.75 | 98.72 | 98.73 | 98.72 |
| 6 | 98.72 | 97.53 | 97.47 | 98.11 | 98.09 |
| 7 | 97.44 | 98.78 | 98.72 | 98.12 | 98.07 |
| 8 | 97.44 | 98.75 | 98.75 | 98.12 | 98.05 |
| 9 | 96.15 | 97.53 | 97.40 | 96.84 | 96.74 |
| 10 | 97.44 | 97.53 | 97.44 | 97.48 | 97.44 |
| Mean ± | 97.44 ± 1.15 | 97.41 ± 1.51 | 97.34 ± 1.49 | 97.42 ± 0.95 | 97.37 ± 0.97 |
Comparison with state-of-the-art approaches.
| PAC-PSO ( | 90.67 | 91.33 | 91.28 | 91.00 | 90.97 |
| HWT ( | 81.71 | 81.43 | 81.48 | 81.57 | 81.60 |
| LR ( | 84.00 | 84.86 | 84.73 | 84.43 | 84.36 |
| CSO ( | 91.84 | 92.40 | 91.92 | 92.13 | 91.88 |
| WRE ( | 93.60 | 93.72 | 93.35 | 93.66 | 93.47 |
| SVM-GA ( | 88.42 | 88.93 | 88.27 | 88.68 | 88.34 |
| LMCoP ( | 89.04 | 90.00 | 89.35 | 89.53 | 89.19 |
| AlexNet (Ours) | 97.44 | 97.41 | 97.34 | 97.42 | 97.37 |
Figure 12Bar plot of comparison of eight algorithms.