| Literature DB >> 35957370 |
Suppat Metarugcheep1, Proadpran Punyabukkana1, Dittaya Wanvarie2, Solaphat Hemrungrojn3,4, Chaipat Chunharas5,6, Ploy N Pratanwanich2,7.
Abstract
Mild cognitive impairment (MCI) is an early stage of cognitive decline or memory loss, commonly found among the elderly. A phonemic verbal fluency (PVF) task is a standard cognitive test that participants are asked to produce words starting with given letters, such as "F" in English and "ก" /k/ in Thai. With state-of-the-art machine learning techniques, features extracted from the PVF data have been widely used to detect MCI. The PVF features, including acoustic features, semantic features, and word grouping, have been studied in many languages but not Thai. However, applying the same PVF feature extraction methods used in English to Thai yields unpleasant results due to different language characteristics. This study performs analytical feature extraction on Thai PVF data to classify MCI patients. In particular, we propose novel approaches to extract features based on phonemic clustering (ability to cluster words by phonemes) and switching (ability to shift between clusters) for the Thai PVF data. The comparison results of the three classifiers revealed that the support vector machine performed the best with an area under the receiver operating characteristic curve (AUC) of 0.733 (N = 100). Furthermore, our implemented guidelines extracted efficient features, which support the machine learning models regarding MCI detection on Thai PVF data.Entities:
Keywords: MoCA; classification; feature extraction; mild cognitive impairment; phonemic clustering; phonemic verbal fluency; silence-based feature; similarity-based feature; switching
Mesh:
Year: 2022 PMID: 35957370 PMCID: PMC9370961 DOI: 10.3390/s22155813
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1Our machine learning framework.
Figure 2The PVF test in the MoCA application. (A) The application will read the PVF instructions “please tell me as many words as possible that begin with the letter “ก” /k/ in one minute” when staff press the speaker button. (B) Space for staff to take notes. (C) Red letters show the timer. PVF, phonemic verbal fluency; MoCA, Montreal cognitive assessment.
Participant demographics.
| MCI (N = 41) | HC (N = 59) | |
|---|---|---|
| Male | 7 | 10 |
| Female | 34 | 49 |
| Word count (mean) | 3–15 (9.61) | 2–24 (10.10) |
| MoCA Score (mean) | 10–24 (21.59) | 25–29 (27) |
MCI, mild cognitive impairment; HC, healthy control; MoCA, Montreal cognitive assessment.
Feature lists.
| Feature | Description |
|---|---|
| Silence-based features | |
| Total silence | Total length of silence during the test. |
| Total voiced | Total length of voiced during the test. |
| Number of silence segments | Total number of silence segments. |
| Number of voice segments | Total number of voice segments. |
| Average silence between word | Total silence divided by the number of silence segments. |
| Q1 Silence | Total silence in the first 30 s of the audio file. |
| Q2 Silence | Total silence in the last 30 s of the audio file. |
| Silence before first word | Silence length before the participant speaks the first word. |
| Different silence between Q1 and Q2 | Total silence in the first 30 s minus the last 30 s. |
| Similarity-based features | |
| Orthographic similarity | Average orthographic similarity value of all words. |
| Levenshtein distance | Average Levenshtein distance ratio of all words. |
| Semantic similarity | Average semantic similarity value of all words. |
| Cluster features | |
| Phonemic clustering | Group of words by phonemic categories. |
| Switching | Total number of the transition between clusters. |
Figure 3Illustration for orthographic similarity. (A) Words are placed at the same index to compare their letters. For calculating the maximum value, each letter in the shorter word is compared with the longest in every index. The quotient is 1/k, where k denotes the overlapped number of words index. The maximum quotients in each letter of the shorter word are summarized and divided by the longer word’s length. (B) The shorter word is shifted by one index; repeat the calculation of the maximum value. (C) Finding the maximum from the values obtained from every lag.
The clusters in Thai.
| Cluster | Characteristic | Example with IPA |
|---|---|---|
| 1 | Word started with “การ” /kaːn/ or “กะ” /kàʔ/ or “กระ” /kràʔ/ | “การเรียน” /kaːn riːan/ “to learn”, “กระต่าย” /kràʔ tàːj/ “rabbit”, “กระโดด” /kràʔ dòːt/ “to jump” |
| 2 | Consonant blends | “กลวง” /kluːaŋ/ “hollow”, “กราบ” /kràːp/ “to pay respects”, “กวาด” /kwàːt/ “to sweep” |
| 3 | Homonym | “ก้าว” /kâːw/ “to step”, “เก้า” /kâːw/ “nine” |
| 4 | Word with only 1 syllable and others | “เกิด” /kə̀ət/ “born”, “แก่” /kɛ̀ɛ/ “old”, “เก็บ” /kèp/ “to store” |
IPA, International Phonetic Alphabet.
Figure 4Illustration for switching.
Classification results for the random forest classifier.
| N | Acc. | F1-Score | Precision | Recall | Specificity | AUC |
|---|---|---|---|---|---|---|
| 1 | 0.584 ± 0.16 | 0.565 ± 0.18 | 0.497 ± 0.24 | 0.535 ± 0.24 | 0.627 ± 0.22 | 0.636 ± 0.20 |
| 2 | 0.584 ± 0.16 | 0.565 ± 0.18 | 0.497 ± 0.24 | 0.535 ± 0.24 | 0.627 ± 0.22 | 0.636 ± 0.20 |
| 3 | 0.584 ± 0.18 | 0.561 ± 0.19 | 0.504 ± 0.26 | 0.530 ± 0.27 | 0.623 ± 0.24 | 0.629 ± 0.21 |
| 4 | 0.574 ± 0.19 | 0.556 ± 0.19 | 0.473 ± 0.21 | 0.510 ± 0.28 | 0.623 ± 0.19 | 0.649 ± 0.20 |
| 5 | 0.534 ± 0.20 | 0.501 ± 0.22 | 0.415 ± 0.29 | 0.375 ± 0.28 | 0.643 ± 0.19 | 0.660 ± 0.23 |
| 6 | 0.594 ± 0.20 | 0.563 ± 0.22 | 0.440 ± 0.26 | 0.450 ± 0.29 | 0.697 ± 0.18 | 0.653 ± 0.22 |
| 7 | 0.590 ± 0.18 | 0.558 ± 0.20 | 0.448 ± 0.26 | 0.500 ± 0.32 | 0.642 ± 0.17 | 0.646 ± 0.22 |
| 8 | 0.640 ± 0.23 * | 0.616 ± 0.25 * | 0.506 ± 0.30 | 0.575 ± 0.37 * | 0.683 ± 0.17 | 0.667 ± 0.23 |
| 9 | 0.610 ± 0.20 | 0.579 ± 0.23 | 0.452 ± 0.28 | 0.550 ± 0.38 | 0.647 ± 0.15 | 0.650 ± 0.19 |
| 10 | 0.580 ± 0.19 | 0.552 ± 0.21 | 0.450 ± 0.28 | 0.455 ± 0.30 | 0.663 ± 0.15 | 0.671 ± 0.21 |
| 11 | 0.620 ± 0.21 | 0.600 ± 0.23 | 0.512 ± 0.30 * | 0.530 ± 0.33 | 0.683 ± 0.17 | 0.683 ± 0.24 * |
| 12 | 0.570 ± 0.18 | 0.545 ± 0.19 | 0.457 ± 0.24 | 0.455 ± 0.27 | 0.647 ± 0.15 | 0.642 ± 0.23 |
| 13 | 0.600 ± 0.17 | 0.565 ± 0.19 | 0.482 ± 0.27 | 0.430 ± 0.26 | 0.717 ± 0.15 * | 0.642 ± 0.25 |
| 14 | 0.580 ± 0.19 | 0.542 ± 0.22 | 0.435 ± 0.32 | 0.430 ± 0.32 | 0.683 ± 0.17 | 0.617 ± 0.22 |
* The maximum value of each feature set; AUC, area under the receiver operating characteristic curve; Acc., accuracy; N, number of selected features, which has highest p-value by chi-square test.
Classification results for the support vector machine classifier.
| N | Acc. | F1-Score | Precision | Recall | Specificity | AUC |
|---|---|---|---|---|---|---|
| 1 | 0.570 ± 0.15 | 0.557 ± 0.15 | 0.494 ± 0.14 | 0.610 ± 0.23* | 0.543 ± 0.21 | 0.665 ± 0.23 |
| 2 | 0.570 ± 0.15 | 0.557 ± 0.15 | 0.494 ± 0.14 | 0.610 ± 0.23* | 0.543 ± 0.21 | 0.669 ± 0.23 |
| 3 | 0.580 ± 0.17 | 0.563 ± 0.17 | 0.490 ± 0.17 | 0.540 ± 0.27 | 0.613 ± 0.18 | 0.672 ± 0.25 |
| 4 | 0.610 ± 0.19 | 0.588 ± 0.20 | 0.515 ± 0.25 | 0.505 ± 0.28 | 0.683 ± 0.17 | 0.680 ± 0.23 |
| 5 | 0.610 ± 0.21 | 0.576 ± 0.22 | 0.523 ± 0.29 | 0.430 ± 0.28 | 0.733 ± 0.20 | 0.717 ± 0.21 |
| 6 | 0.650 ± 0.22 * | 0.626 ± 0.24 * | 0.567 ± 0.28 | 0.525 ± 0.31 | 0.733 ± 0.20 | 0.721 ± 0.21 |
| 7 | 0.650 ± 0.21 * | 0.624 ± 0.22 | 0.583 ± 0.28 * | 0.505 ± 0.28 | 0.750 ± 0.20 | 0.729 ± 0.20 |
| 8 | 0.590 ± 0.18 | 0.551 ± 0.18 | 0.539 ± 0.29 | 0.365 ± 0.20 | 0.750 ± 0.20 | 0.725 ± 0.21 |
| 9 | 0.530 ± 0.11 | 0.362 ± 0.09 | 0.200 ± 0.40 | 0.025 ± 0.08 | 0.883 ± 0.17 | 0.733 ± 0.20 * |
| 10 | 0.540 ± 0.11 | 0.366 ± 0.09 | 0.250 ± 0.43 | 0.025 ± 0.07 | 0.900 ± 0.17 | 0.733 ± 0.20 |
| 11 | 0.550 ± 0.07 | 0.356 ± 0.03 | 0.000 ± 0.00 | 0.000 ± 0.00 | 0.933 ± 0.11 | 0.733 ± 0.20 |
| 12 | 0.560 ± 0.07 | 0.358 ± 0.03 | 0.000 ± 0.00 | 0.000 ± 0.00 | 0.950 ± 0.11 * | 0.725 ± 0.21 |
| 13 | 0.560 ± 0.07 | 0.358 ± 0.03 | 0.000 ± 0.00 | 0.000 ± 0.00 | 0.950 ± 0.11 * | 0.725 ± 0.21 |
| 14 | 0.560 ± 0.07 | 0.358 ± 0.03 | 0.000 ± 0.00 | 0.000 ± 0.00 | 0.950 ± 0.11 * | 0.725 ± 0.21 |
* The maximum value of each feature set; AUC, area under the receiver operating characteristic curve; Acc., accuracy; N, number of selected features, which has highest p-value by chi-square test.
Classification results for the XGBoost classifier.
| N | Acc. | F1-Score | Precision | Recall | Specificity | AUC |
|---|---|---|---|---|---|---|
| 1 | 0.620 ± 0.15 | 0.594 ± 0.17 | 0.521 ± 0.24 | 0.605 ± 0.28 * | 0.633 ± 0.22 | 0.640 ± 0.21 |
| 2 | 0.620 ± 0.15 | 0.594 ± 0.17 | 0.521 ± 0.24 | 0.605 ± 0.28 * | 0.633 ± 0.22 | 0.640 ± 0.21 |
| 3 | 0.590 ± 0.17 | 0.558 ± 0.19 | 0.475 ± 0.24 | 0.480 ± 0.27 | 0.663 ± 0.19 | 0.626 ± 0.17 |
| 4 | 0.560 ± 0.14 | 0.515 ± 0.15 | 0.438 ± 0.21 | 0.390 ± 0.23 | 0.680 ± 0.20 | 0.659 ± 0.13 |
| 5 | 0.550 ± 0.17 | 0.497 ± 0.20 | 0.343 ± 0.28 | 0.400 ± 0.34 | 0.647 ± 0.19 | 0.550 ± 0.23 |
| 6 | 0.570 ± 0.17 | 0.540 ± 0.19 | 0.447 ± 0.24 | 0.480 ± 0.27 | 0.630 ± 0.21 | 0.638 ± 0.19 |
| 7 | 0.560 ± 0.17 | 0.526 ± 0.19 | 0.433 ± 0.22 | 0.450 ± 0.29 | 0.630 ± 0.19 | 0.617 ± 0.17 |
| 8 | 0.590 ± 0.20 | 0.564 ± 0.21 | 0.489 ± 0.23 | 0.500 ± 0.30 | 0.650 ± 0.22 | 0.621 ± 0.21 |
| 9 | 0.570 ± 0.13 | 0.536 ± 0.16 | 0.420 ± 0.22 | 0.455 ± 0.28 | 0.647 ± 0.13 | 0.638 ± 0.18 |
| 10 | 0.580 ± 0.14 | 0.550 ± 0.16 | 0.437 ± 0.22 | 0.480 ± 0.27 | 0.647 ± 0.13 | 0.642 ± 0.17 |
| 11 | 0.630 ± 0.11 * | 0.603 ± 0.13 * | 0.522 ± 0.16 * | 0.530 ± 0.26 | 0.697 ± 0.09 | 0.642 ± 0.23 |
| 12 | 0.630 ± 0.18 * | 0.585 ± 0.22 | 0.522 ± 0.35 * | 0.455 ± 0.34 | 0.747 ± 0.15 * | 0.650 ± 0.25 |
| 13 | 0.620 ± 0.17 | 0.592 ± 0.17 | 0.512 ± 0.24 | 0.505 ± 0.25 | 0.697 ± 0.14 | 0.671 ± 0.18 * |
| 14 | 0.630 ± 0.13 * | 0.601 ± 0.16 | 0.513 ± 0.23 | 0.505 ± 0.25 | 0.713 ± 0.10 | 0.629 ± 0.23 |
* The maximum value of each feature set; AUC, area under the receiver operating characteristic curve; Acc., accuracy; N, number of selected features, which has highest p-value by chi-square test.
Figure 5Feature importance explained by the SHAP value for the random forest classifier.
Figure 6Feature importance explained by the SHAP value for the XGBoost classifier.
Figure 7Feature importance explained by the SHAP value for the SVM classifier.
Figure 8The p-values obtained from the chi-square test of the feature-selection process.
Figure 9Distribution of the feature values between MCI and HC. Green triangles represent the data means. The orange lines show the medians of the data. White circles are the data outliers. MCI, mild cognitive impairment; HC, healthy control.