| Literature DB >> 35551276 |
Eiron John Lugtu1, Denise Bernadette Ramos1, Alliah Jen Agpalza1, Erika Antoinette Cabral1, Rian Paolo Carandang1, Jennica Elia Dee1, Angelica Martinez1, Julius Eleazar Jose1, Abegail Santillan2,3, Ruth Bangaoil2,3,4, Pia Marie Albano2,3,5, Rock Christian Tomas6.
Abstract
Given the increasing prevalence of lung cancer worldwide, an auxiliary diagnostic method is needed alongside the microscopic examination of biopsy samples, which is dependent on the skills and experience of pathologists. Thus, this study aimed to advance lung cancer diagnosis by developing five (5) artificial neural network (NN) models that can discriminate malignant from benign samples based on infrared spectral data of lung tumors (n = 122; 56 malignant, 66 benign). NNs were benchmarked with classical machine learning (CML) models. Stratified 10-fold cross-validation was performed to evaluate the NN models, and the performance metrics-area under the curve (AUC), accuracy (ACC) positive predictive value (PPV), negative predictive value (NPV), specificity rate (SR), and recall rate (RR)-were averaged for comparison. All NNs were able to outperform the CML models, however, support vector machine is relatively comparable to NNs. Among the NNs, CNN performed best with an AUC of 92.28% ± 7.36%, ACC of 98.45% ± 1.72%, PPV of 96.62% ± 2.30%, NPV of 90.50% ± 11.92%, SR of 96.01% ± 3.09%, and RR of 89.21% ± 12.93%. In conclusion, NNs can be potentially used as a computational tool in lung cancer diagnosis based on infrared spectroscopy of lung tissues.Entities:
Mesh:
Year: 2022 PMID: 35551276 PMCID: PMC9098097 DOI: 10.1371/journal.pone.0268329
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.752
Fig 1Experimental design process flowchart.
The figure shows the experimental design of the study from acquisition of spectral data to machine learning training and evaluation.
Fig 2FNN design architecture.
Fig 3RNN design architecture.
Fig 4CNN design architecture.
Neural network hyperparameters for GA.
| GA-optimized hyperparameters | |||||
|---|---|---|---|---|---|
| FNN (N = 2, N = 4, N = 8) | RNN | CNN | |||
| variable | search space | variable | search space | variable | search space |
| Epoch ( | [10,300] ∈ ℕ | Epoch ( | [10,300] ∈ ℕ | Epoch ( | [10,300] ∈ ℕ |
| Learning rate ( | 10- | Learning rate ( | 10- | Learning rate ( | 10- |
| Neurons per layer ( | [2,30] ∈ ℕ | RNN input partitions ( | [10,100] ∈ ℕ | Number of conv. Layers ( | [1,5] ∈ ℕ |
| AdaGrad optimization constant (ε) | 10- | Neurons per fold ( | [2,30] ∈ ℕ | Filter size ( | [2,11] ∈ ℕ |
| Neurons per FNN layer ( | [2,30] ∈ ℕ | Filter skip ( | [2, | ||
| AdaGrad optimization constant ( | 10- | Kernel count for N = 1 ( | 2 | ||
| Neurons per FNN layer ( | [2,30] ∈ ℕ | ||||
| AdaGrad optimization constant ( | 10- | ||||
|
| |||||
Hyperparameters of GA design.
| Hyperparameter | Description/Value |
|---|---|
| Maximum number of generations ( | 30 |
| Number of individuals ( | 30 |
| Mutation rate (% | 0% for the fittest 50%; else 5% |
| Crossover rate (% | 100% for fittest 50%; else 0% |
| Crossover method | Single point crossover |
| Fitness function | Validation set accuracy |
| Elitism | Fittest 50% as parents with 100% survival rate, and 50% as new individuals |
| Termination criteria | Generation count reaches |
Fig 5Median ATR-FTIR absorbance spectra of malignant (n = 53) and benign (n = 65) lung tissue samples.
The figure shows the median FTIR spectrum of malignant and benign lung tissue samples and their corresponding peaks identified via visual analysis.
Computation of the spectrum variables (peak positions and normalized absorbances) of malignant and benign lung samples in the fingerprint IR region (1800 cm-1 to 850 cm-1).
| Malignant Samples (n = 53) | Benign Samples (n = 65) | Functional Group | Vibrational Mode | Molecular Source [ | |||
|---|---|---|---|---|---|---|---|
| Peak Position | Mean abs ± SD | Peak Position | Mean abs ± SD | ||||
| 1636 | 0.9605 ± 0.0856 | 1638 | 0.9885± 0.0139 | O = C–N–H | Amide I, protein | 0.6991 | |
|
|
|
|
|
|
|
|
|
| 1452 | 0.3659 ± 0.0687 | 1452 | 0.3714± 0.0510 | –(CH2)n,–(CH3)n– | δas(CH3), δas(CH2), δs(CH3) | Lipids | 0.9851 |
| 1401 | 0.3244 ± 0.0694 | 1397 | 0.3480 ± 0.0542 | –(CH2)n– | δs(CH3) | Lipids | 0.0590 |
|
|
|
|
|
|
|
|
|
| 1236 | 0.165 ± 0.0766 | 1236 | 0.1921 ± 0.0775 | RO–PO2−–OR | DNA, RNA, phospholipids | 0.1138 | |
| 1160 | 0.0467 ± 0.0343 | 1160 | 0.0506 ± 0.0475 | C–O–H | ν(CO), γ(COH) | Carbohydrates | 0.4302 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
* Mann-Whitney U test (two-tailed); significant when p<0.05.
**Values in bold refer to significantly different peak absorbance between malignant and benign samples (p>0.05).
Abbreviations: v: stretching; δ: bending; γ: wagging, twisting and rocking; s: symmetric; as: asymmetric; def: deformation.
Fig 6PCA biplot showing the distribution of malignant and benign samples and the variances contributed by each biomolecule.
The red points represent the malignant samples while the blue points represent the benign samples.
Fig 7Average performance accuracy of NN models per generation.
The plots show the average accuracy of each NN model during the GA-based NN hyperparameter tuning process. The averaged metric shown for each generation is derived from the metric of the best individual over 50 trials. Evident in the GA plots, the FNN models were the fastest to achieve steady-state performance while the RNN model was the slowest. The RNN plot also shows a comparatively larger range of values per generation, which may suggest that the search space for RNN models of very high accuracy is relatively smaller than those of the FNNs and the CNN; hence RNN models may be the most difficult to tune. A. Average performance accuracy of FNN2-type individuals per generation. B. Average performance accuracy of FNN4-type individuals per generation. C. Average performance accuracy of FNN8-type individuals per generation. D. Average performance accuracy of CNN-type individuals per generation. E. Average performance accuracy of RNN-type individuals per generation.
FNN hyperparameters.
| Median (25th percentile, 75th percentile) | |||
|---|---|---|---|
| Epoch | Learning Rate | Neurons per Layer | |
|
| 241 (201.5, 268) | 3.95 × 10−3 (1.83 ×10−3, 1.59 × 10−2) | 17 (7.75, 24) |
|
| 229.5 (166, 267) | 2.17 × 10−3 (1.60 × 10−3, 5.01 ×10−3) | 19 (15.75, 25.5) |
|
| 218.5 (189, 270.25) | 1.75 × 10−3 (1.13 × 10−3, 3.45 × 10−3) | 18 (13, 26) |
CNN hyperparameters.
| Epoch | Learning Rate | Filter Size | Skip Length | Number of Conv. Layers (N) | Kernel size at N = 1 | FNN neurons per layer | |
|---|---|---|---|---|---|---|---|
|
| 205 | 1.07 ×10−3 | 8 | 4 | 2 | 8 | 17.5 |
| 162 | 1.04 × 10−5 | 6 | 2.75 | 1 | 7 | 14.75 | |
| 221.25 | 2.02 × 10−2 | 9 | 4 | 2 | 16 | 22 |
RNN hyperparameters.
| Epoch | Learning Rate | RNN Neurons per Fold | FNN Neurons per Layer | |
|---|---|---|---|---|
|
| 202.5 | 1.43 × 10−3 | 24 | 21 |
| 152.5 | 3.77 ×10−5 | 21 | 17.75 | |
| 264.25 | 8.60 × 10−2 | 28.25 | 26.25 |
Mean and standard deviation of diagnostic performance of all the machine learning models.
| FNN 2 | FNN 4 | FNN 8 | RNN | CNN | DT | RF | NB | LDA | LR | SVM | |
|---|---|---|---|---|---|---|---|---|---|---|---|
|
| 92.41 ± 10.28 | 92.85 ± 9.98 | 92.77 ± 9.62 | 90.40 ± 11.62 | 92.28 ± 7.36 | 78.93 ± 19.87 | 92.15 ± 13.79 | 77.91 ± 21.22 | 62.92 ± 13.44 | 82.16 ± 19.84 | 99.38 ± 1.97 |
|
| 98.41 ± 4.07 | 98.39 ± 3.57 | 97.61 ± 4.91 | 95.98 ± 6.25 | 98.45 ± 1.72 | 77.58 ± 16.91 | 85.87 ± 15.11 | 75.13 ± 19.14 | 65.45 ± 14.60 | 72.19 ± 18.68 | 94.38 ± 9.69 |
|
| 94.92 ± 9.49 | 96.03 ± 8.63 | 96.48 ± 8.01 | 90.91 ± 11.99 | 96.62 ± 2.30 | 73.95 ± 30.28 | 83.55 ± 26.14 | 64.65 ± 33.42 | 59.54± 20.84 | 65.25 ± 36.13 | 93.85 ± 14.92 |
|
| 92.79 ± 11.95 | 93.57 ± 11.28 | 93.22 ± 10.75 | 94.03 ± 9.26 | 90.50 ± 11.92 | 80.11 ± 21.98 | 87.60 ± 18.36 | 82.21 ± 22.01 | 70.25 ± 15.13 | 94.89 ± 30.57 | 94.57 ± 12.11 |
|
| 94.60 ± 11.41 | 94.67 ± 11.57 | 95.30 ± 10.64 | 86.63 ± 17.58 | 96.01 ± 3.09 | 84.36 ± 18.04 | 90.78 ± 14.34 | 79.79 ± 18.68 | 69.08 ± 14.50 | 74.97 ± 25.98 | 96.75 ± 8.03 |
|
| 89.84 ± 16.36 | 90.88 ± 16.13 | 90.10 ± 15.66 | 92.87 ± 13.18 | 89.21 ± 12.93 | 73.35 ± 28.25 | 83.74 ± 24.06 | 71.10 ± 33.83 | 62.13± 17.14 | 64.49 ± 35.74 | 94.46 ± 12.64 |
Difference of average performance metric between NN models and the SVM model.
| Difference of average performance metric ( | ||||||
|---|---|---|---|---|---|---|
| AUC(%) | ACC(%) | PPV(%) | NPV(%) | SR(%) | RR(%) | |
|
|
|
| 1.07 (0.1094) |
|
|
|
|
|
|
|
| -1.01 (0.4156) | -2.08 (0.0734) |
|
|
|
|
|
| -1.36 (0.9560) | -1.45 (0.1992) |
|
|
|
|
|
| -0.54 (0.8857) |
| -1.59 (0.0713) |
|
|
|
|
|
| -0.74 (0.8493) |
|
Fig 8Distribution of discordant samples from concordant samples via PCA.
The plot shows the diagnosis of the models per discordant samples (dark-colored points) over the distribution of concordant malignant samples (light red) and concordant benign (light blue) samples. The diagnoses of the models for all discordant samples were consistent with the original diagnoses of the study sites.
Fig 9Prediction probability of NN models per discordant samples.
The figures show the prediction probability of each NN model for the discordant benign (n = 13) and discordant malignant (n = 11) samples. The discordant samples were grouped according to the diagnosis by the pathologist of their respective study sites. All NN models show a median prediction score that is above the 0.5 (50%) mark, meaning that all the models had the same prediction as that of the diagnosis of the pathologist. A. Prediction probability of FNN2 models per discordant samples. B. Prediction probability of FNN4 models per discordant samples. C. Prediction probability of FNN8 models per discordant samples. D. Prediction probability of CNN models per discordant samples. E. Prediction probability of RNN models per discordant samples.