| Literature DB >> 34384201 |
Ilias Tougui1, Abdelilah Jilbab1, Jamal El Mhamdi1.
Abstract
OBJECTIVE: With advances in data availability and computing capabilities, artificial intelligence and machine learning technologies have evolved rapidly in recent years. Researchers have taken advantage of these developments in healthcare informatics and created reliable tools to predict or classify diseases using machine learning-based algorithms. To correctly quantify the performance of those algorithms, the standard approach is to use cross-validation, where the algorithm is trained on a training set, and its performance is measured on a validation set. Both datasets should be subject-independent to simulate the expected behavior of a clinical study. This study compares two cross-validation strategies, the subject-wise and the record-wise techniques; the subject-wise strategy correctly mimics the process of a clinical study, while the record-wise strategy does not.Entities:
Keywords: Data Analysis; Diagnosis; Machine Learning; Parkinson Disease; Statistical Models
Year: 2021 PMID: 34384201 PMCID: PMC8369053 DOI: 10.4258/hir.2021.27.3.189
Source DB: PubMed Journal: Healthc Inform Res ISSN: 2093-3681
Final distribution of valid subjects in this study
| PD group | HC group | Total | |
|---|---|---|---|
| Number of recordings | 424 | 424 | 848 |
|
| |||
| Number of subjects | 212 | 212 | 424 |
|
| |||
| Sex | |||
| Male | 161 | 161 | 322 |
| Female | 51 | 51 | 102 |
|
| |||
| Age (yr) | 58.97±8.95 (40–79) | 58.97±8.95 (40–79) | |
Values are presented as mean ± standard (min–max).
PD: Parkinson’s disease, HC: healthy controls.
Figure 1Subject-wise and record-wise divisions.
Record-wise and subject-wise cross-validation (CV) techniques
| CV group | CV technique | Description |
|---|---|---|
| Record-wise group | Stratified k-folds CV (skfcv) | With this technique, the dataset is divided into k blocks (folds) in a stratified manner [ |
| Leave-one-out CV (loocv) | In this technique, only one record is left out for each learning process [ | |
| Repeated stratified k-folds CV (rskfcv) | This technique is similar to stratified k-folds CV, but it is repeated n times [ | |
|
| ||
| Subject-wise group | Stratified-group k-folds CV (sgkfcv) | Using this technique, the dataset is divided into k blocks in a stratified manner with group of subjects [ |
| Leave-one-group-out CV (logocv) | In this technique, we leave out the records of only one group of subjects for each learning process [ | |
| Repeated stratified-group k-folds CV (rsgkfcv) | This technique is similar to stratified-group-k-folds CV, but it is repeated n times [ | |
Support vector machine (SVM) and random forest (RF) pipelines with their hyperparameters
| SVM pipeline | Hyperparameters | RF pipeline | Hyperparameters |
|---|---|---|---|
| SVM_Pipeline = { | Imputation = { | RF_Pipeline = { | Imputation = { |
Performance measures
| Performance measure | Equation | Description |
|---|---|---|
| Accuracy |
| Accuracy is the proportion of correctly classified participants diagnosed with and without Parkinson’s disease among the total number of cases examined. |
| Sensitivity |
| Sensitivity measures the proportion of participants diagnosed with Parkinson’s disease who have been correctly identified by the classifier. |
| Specificity |
| Specificity measures the proportion of healthy participants who are correctly identified by the classifier. |
| F1 score |
| The F1 score is a measure of the accuracy of a test. It is calculated from the precision and the sensitivity of the test. |
The following definitions are used in the equations:
True positive (TP) refers to the number of participants diagnosed with Parkinson’s disease who are correctly identified by the classifier.
True negative (TN) denotes the number of healthy participants who are correctly identified by the classifier.
False positive (FP) refers to the number of healthy participants who are incorrectly identified by the classifier.
False negative (FN) denotes the number of participants diagnosed with Parkinson’s disease who are incorrectly diagnosed by the classifier.
Figure 2Performance of the support vector machine (SVM) pipeline with various cross-validation (CV) techniques compared to subject- wise division. (A) Accuracy. (B) Sensitivity. (C) Specificity. (D) F1 score.
Figure 3Performance of the random forest (RF) pipeline with various cross-validation (CV) techniques compared to subject-wise division. (A) Accuracy. (B) Sensitivity. (C) Specificity. (D) F1 score.
Figure 4Performance of the support vector machine (SVM) pipeline with various cross-validation (CV) techniques compared to record-wise division. (A) Accuracy. (B) Sensitivity. (C) Specificity. (D) F1 score.
Figure 5Performance of the random forest (RF) pipeline with various cross-validation (CV) techniques compared to record-wise division. (A) Accuracy. (B) Sensitivity. (C) Specificity. (D) F1 score.
Execution times of the different cross-validation (CV) techniques
| CV group | CV techniques | Execution time | Is the technique valid in this study? |
|---|---|---|---|
| Record-wise | Stratified k-folds CV | 1 min | No |
| Leave-one-out CV | 6 hr 16 min 2 s | No | |
| Repeated stratified k-folds CV | 4 min 49 s | No | |
|
| |||
| Subject-wise | Stratified-group k-folds CV | 1 min 10 s | Yes |
| Leave-one-group-out CV | 31 min 35 s | Yes | |
| Repeated stratified-group k-folds CV | 4 min 59 s | Yes | |