| Literature DB >> 36091110 |
Kai Wang1, Peizhe Chen2, Bojian Feng3,4, Jing Tu1, Zhengbiao Hu1, Maoliang Zhang1, Jie Yang1, Ying Zhan1, Jincao Yao3,4,5, Dong Xu3,4,5,6.
Abstract
Objective: To build a machine learning (ML) prediction model for prostate cancer (PCa) from transrectal ultrasound video clips of the whole prostate gland, diagnostic performance was compared with magnetic resonance imaging (MRI).Entities:
Keywords: artificial intelligence; machine learning; prostate cancer; support vector machine; ultrasound
Year: 2022 PMID: 36091110 PMCID: PMC9459141 DOI: 10.3389/fonc.2022.948662
Source DB: PubMed Journal: Front Oncol ISSN: 2234-943X Impact factor: 5.738
Figure 1The flowchart of inclusion and exclusion of the study population.
Figure 2Overall flow chart of the study, including image acquisition and segmentation, feature extraction, feature selection, machine learning, and evaluation.
Characteristics of patients in the training, validation and test datasets.
| Training set | Validation set | Test set |
| |
|---|---|---|---|---|
| Age(y)* | 72.02±8.721 | 71.21±6.246 | 69.64±8.262 | 0.161 |
| PSA(ng/mL)* | 19.91±44.56 | 22.77±65.51 | 46.98±114.89 | 0.065 |
| Pathology | 0.871 | |||
| No.of Benign(-)(%) | 83(48.8%) | 15(44.1%) | 30(49.2%) | |
| BPH | 68(40%) | 13(38.3%) | 25(41%) | |
| BPH & prostatitis | 11(6.4%) | 1(2.9%) | 3(4.9%) | |
| BPH & BCH | 1(0.6%) | 0 | 2(3.2) | |
| BPH & LGIN | 3(1.8%) | 1(2.9%) | 0 | |
| No. of Pca(+)(%) | 87(51.2%) | 19(55.9%) | 31(50.8%) | |
| GS6 | 30(17,6%) | 9(26.5%) | 14(23%) | |
| GS7 | 38(22.4%) | 7(20.6%) | 8(13.1%) | |
| GS8 | 10(5.9%) | 2(5.9%) | 6(9.8%) | |
| GS>=8 | 9(5.3%) | 1(2.9%) | 3(4.9%) |
BPH, benign prostatic hyperplasia; BCH, basal cell hyperplasia; LGIN. low-grade intraepithelial neoplasia.*Data are expressed as mean ± standard deviation.p< 0.05 indicates significant differences in patients’ clinicopathological features in the validation and test sets.
The subset of radiomics features ultimately selected by the LASSO algorithm.
| Feature | Image type | Feature Class | Feature Name | LASSO coefficients |
|---|---|---|---|---|
| 1 | Original | Firstorder | Range | 0.049561 |
| 2 | Original | glcm | ClusterProminence | 0.004193 |
| 3 | Original | glszm | ZoneEntropy | 0.011709 |
| 4 | Wavelet-LHL | firstorder | Skewness | -0.055034 |
| 5 | Wavelet-LHL | glcm | ClusterShade | -0.025479 |
| 6 | Wavelet-LHL | glcm | Correlation | 0.010685 |
| 7 | Wavelet-LHH | gldm | LargeDependenceLowGrayLevelEmphasis | 0.018646 |
| 8 | Wavelet-HLL | glszm | GrayLevelNonUniformity | -0.073279 |
| 9 | Wavelet-HHH | firstorder | Median | -0.050800 |
| 10 | Wavelet-HHH | glcm | ClusterShade | 0.025124 |
| 11 | Wavelet-HHH | gldm | LargeDependenceLowGrayLevelEmphasis | 0.039665 |
| 12 | Wavelet-LLL | glszm | LargeAreaHighGrayLevelEmphasis | -0.026230 |
| 13 | Wavelet-LLL | glszm | SizeZoneNonUniformityNormalized | 0.021580 |
| 14 | Wavelet-LLL | glszm | SmallAreaHighGrayLevelEmphasis | 0.056598 |
First-order features describe the distribution of voxel intensities within the image region defined by the mask through commonly used and basic metrics. GLCM features describe the second-order joint probability function of an image region constrained by the mask. They are defined as P. GLDM features quantify gray-level dependencies in an image, and GLSZM features quantify gray-level zones in an image.
Figure 3Selection of significant parameters in features in the training set and definition of the linear predictor. (A) Spearman’s correlation coefficients were calculated for the fourteen selected features. (B) Characters classification weight of the features.
Figure 4Generation of the optimal penalization coefficient lambda. (A) Ten-time cross-validation for tuning parameter selection in the LASSO model. (B) LASSO coefficient solution path for the 14 features.
Diagnostic performance of machine learning model and MRI on a per-lesion basis.
| Dataset and Method | Sensitivity (95% CI) | Specificity (95% CI) | Precision (95% CI) | AUC |
| Kappa |
|---|---|---|---|---|---|---|
|
| ||||||
|
| 0.63 (0.38-0.83) | 0.80 (0.51-0.95) | 0.80 (0.51-0.95) | 0.78 | 0.012 | 0.42 |
|
| 0.65 (0.45-0.80) | 0.67 (0.47-0.82) | 0.67 (0.47-0.82) | 0.75 | 0.015 | 0.312 |
|
| ||||||
|
| 0.63 (0.37-0.83) | 0.87 (0.58-0.98) | 0.86 (0.56-0.97) | 0.77 | 0.003 | 0.481 |
|
| 0.45 (0.28-0.64) | 0.93 (0.76-0.99) | 0.88 (0.60-0.98) | 0.69 | 0.001 | 0.382 |
|
| ||||||
|
| 0.63 (0.39-0.83) | 0.67 (0.39-0.87) | 0.71 (0.44-0.87) | 0.65 | 0.084 | 0.294 |
|
| 0.71 (0.52-0.85) | 0.60 (0.41-0.77) | 0.65 (0.46-0.80) | 0.65 | 0.015 | 0.31 |
|
| ||||||
|
| 0.63 (0.39-0.83) | 0.87 (0.58-0.98) | 0.86 (0.56-0.97) | 0.75 | 0.003 | 0.481 |
|
| 0.61 (0.42-0.78) | 0.83 (0.65-0.93) | 0.79 (0.57-0.92) | 0.72 | 0.0003 | 0.445 |
|
| ||||||
|
| 0.74 (0.49-0.90) | 0.87 (0.58-0.98) | 0.88 (0.60-0.98) | 0.8 | 0.000464 | 0.591 |
|
| 0.68 (0.49-0.83) | 0.70 (0.50-0.85) | 0.70 (0.50-0.85) | 0.72 | 0.003 | 0.377 |
|
| ||||||
|
| 0.74 (0.49-0.90) | 0.93 (0.66-0.99) | 0.93 (0.66-0.99) | 0.85 | 0.00009 | 0.652 |
|
| 0.81 (0.62-0.92) | 0.80 (0.61-0.92) | 0.81 (0.62-0.92) | 0.81 | 0.000002 | 0.606 |
SVM model, support vector machine model; RF model, random forest model; MRI-JR, junior radiologists’ (less than 5 years of experience) diagnosis based on MRI; MRI-SR, senior radiologists’ (more than 5 years of experience) diagnosis based on MRI. p< 0.05 indicates a significant difference in the discrimination of the SVM model and MRI diagnosis.
Figure 5Comparison of ROC between the ML models and MRI in the validation set and test set. (A) shows the ROC curves of the validation set. (B) shows the ROC curves of the test set. (MRI-JR: junior radiologists’ diagnosis based on MRI, MRI-SR: senior radiologists’ diagnosis based on MRI).