| Literature DB >> 28465706 |
Dan Meng1, Guitao Cao1,2, Ye Duan2, Minghua Zhu1, Liping Tu2,3, Dong Xu2, Jiatuo Xu4.
Abstract
Computer aided tongue diagnosis has a great potential to play important roles in traditional Chinese medicine (TCM). However, the majority of the existing tongue image analyses and classification methods are based on the low-level features, which may not provide a holistic view of the tongue. Inspired by deep convolutional neural network (CNN), we propose a novel feature extraction framework called constrained high dispersal neural networks (CHDNet) to extract unbiased features and reduce human labor for tongue diagnosis in TCM. Previous CNN models have mostly focused on learning convolutional filters and adapting weights between them, but these models have two major issues: redundancy and insufficient capability in handling unbalanced sample distribution. We introduce high dispersal and local response normalization operation to address the issue of redundancy. We also add multiscale feature analysis to avoid the problem of sensitivity to deformation. Our proposed CHDNet learns high-level features and provides more classification information during training time, which may result in higher accuracy when predicting testing samples. We tested the proposed method on a set of 267 gastritis patients and a control group of 48 healthy volunteers. Test results show that CHDNet is a promising method in tongue image classification for the TCM study.Entities:
Year: 2017 PMID: 28465706 PMCID: PMC5390589 DOI: 10.1155/2017/7452427
Source DB: PubMed Journal: Evid Based Complement Alternat Med ISSN: 1741-427X Impact factor: 2.629
Figure 1Flowchart of normal and abnormal detection framework based on CHDNet.
Algorithm 1Tongue images classification based on CHDNet.
Figure 2The structure of the two-stage CHDNet.
Figure 3Illustration of 2 × 2 patch size taking for a 5 × 5 image.
Performance comparison on the tongue images dataset with the proposed components of CHDNet.
| Method | Accuracy | Sensitivity | Specificity |
|---|---|---|---|
| PCANet | 84.77% | 100.00% | 0.00% |
| PCANet + NT | 85.40% | 100.00% | 4.17% |
| PCANet + MFA | 86.01% | 99.25% | 12.31% |
| PCANet + HD | 87.37% | 98.16% | 27.35% |
| PCANet + LRN | 84.77% | 100.00% | 0.00% |
| CHDNet with all four components | 91.14% | 94.26% | 75.40% |
Performance comparison for weighted LIBLINEAR SVM.
| Normal : Abnormal | Accuracy | Sensitivity | Specificity |
|
|---|---|---|---|---|
| | 90.61% | 93.79% | 72.93% | 68.40% |
| | 91.01% | 94.22% | 73.36% | 69.12% |
| | 90.62% | 94.57% | 68.60% | 64.88% |
| | 90.94% | 94.19% | 72.89% | 68.66% |
| | 91.02% | 94.79% | 70.07% | 66.42% |
| | 90.55% | 94.28% | 69.89% | 66.25% |
| | 90.59% | 94.87% | 66.91% | 63.48% |
| | 91.14% | 94.26% | 75.40% | 71.07% |
| | 91.04% | 94.39% | 72.27% | 68.22% |
Some normal and abnormal tongue images classified by our method.
| Patient's number | Original image | Mask | Tongue body | Normalization | Predict label | Actual label |
|---|---|---|---|---|---|---|
| N017 |
|
|
|
| 0 | 0 |
| N018 |
|
|
|
| 0 | 0 |
| D100 |
|
|
|
| 1 | 1 |
| X084 |
|
|
|
| 1 | 1 |
The first column is the original image of 1728 × 1296 pixels; the second column is the background mask; the third column is the extraction of tongue body; the fourth column is the color space transformed tongue body image; the fifth column is the normalized 512 × 512 pixels image; the last column is the predicted label.
Pathological information by TCM.
| Patient's number | Pathological feature | Clinical practitioner's subjective diagnosis | Diagnosis result |
|---|---|---|---|
| N017 | — | — | Normal |
|
| |||
| N018 | — | — | Normal |
|
| |||
| D100 | Superficial gastritis | Cold Zheng-deficiency cold of the spleen | Abnormal |
|
| |||
| X084 | Atrophic gastritis | Hot Zheng-damp heat in the spleen and the stomach | Abnormal |
The first column is the patient's number, the second column is pathological feature, the third column is clinical practitioners subjective diagnosis, and the last column is the clinical practitioners diagnosis result.
Comparison of the proposed method with other feature extracting approaches.
| LDA | KNN | CART | GBDT | RF | LIBSVM | LIBLEAR SVM | ||
|---|---|---|---|---|---|---|---|---|
| HOG [ | Sensitivity | 91.65% | 100.00% | 90.19% | 100.00% | 100.00% | 99.93% | 98.02% |
| Specificity | 58.00% | 22.33% | 56.31% | 54.58% | 47.42% | 48.82% | 51.27% | |
| LBP | Sensitivity | 99.70% | 100.00% | 90.49% | 92.06% | 100.00% | 100.00% | 100.00% |
| Specificity | 64.56% | 60.93% | 65.91% | 65.16% | 58.29% | 58.64% | 58.84% | |
| SIFT | Sensitivity | 98.95% | 100.00% | 90.49% | 92.78% | 100.00% | 98.28% | 98.17% |
| Specificity | 41.44% | 0.20% | 58.18% | 55.20% | 62.31% | 44.53% | 45.42% | |
| HOG + LBP | Sensitivity | 99.93% | 99.96% | 91.38% | 92.85% | 100.00% | 100.00% | 100.00% |
| Specificity | 49.60% | 56.69% | 64.49% | 67.24% | 58.20% | 60.33% | 60.36% | |
| HOG + SIFT | Sensitivity | 100.00% | 100.00% | 91.87% | 92.32% | 100.00% | 98.19% | 97.76% |
| Specificity | 59.91% | 0.42% | 60.56% | 58.29% | 62.49% | 43.87% | 46.62% | |
| LBP + SIFT | Sensitivity | 100.00% | 100.00% | 91.58% | 91.90% | 100.00% | 98.31% | 97.87% |
| Specificity | 58.89% | 0.82 | 62.31% | 63.33% | 60.20% | 44.07% | 44.33% | |
| HOG + LBP + SIFT | Sensitivity | 99.96% | 100.00% | 91.95% | 92.85% | 100.00% | 98.27% | 98.24% |
| Specificity | 59.67% | 0.87% | 65.64% | 66.33% | 59.93% | 43.18% | 46.18% | |
| Doublets [ | Sensitivity | 91.88% | 100.00% | 93.22% | 93.67% | 100.00% | 100.00% | 100.00% |
| Specificity | 36.71% | 0.00% | 48.91% | 49.42 | 25.51% | 0.00% | 0.00% | |
| Doublets + HOG [ | Sensitivity | 92.35% | 100.00% | 94.22% | 94.38% | 100.00% | 100.00% | 100.00% |
| Specificity | 29.96% | 0.00% | 51.80% | 46.44% | 27.29% | 0.00% | 0.00% | |
| PCANet [ | Sensitivity | 100.00% | 98.35% | 87.05% | 88.50% | 100.00% | 100.00% | 100.00% |
| Specificity | 0.00% | 14.09% | 29.51% | 28.40 | 0.00% | 0.00% | 0.00% | |
| Our method | Sensitivity | 90.68% | 93.11% | 92.44% | 92.63% | 93.37% | 94.68% | 94.26% |
| Specificity | 52.91% | 69.18% | 61.56% | 59.98% | 65.62% | 71.18% | 75.40% |
Note. The sum rule of feature combination is a cascade operation. Given two types of features f and f obtained by feature extraction methods FEM and FEM, respectively, then FEM + FEM is equal to [f, f]. For example, HOG + LBP means, for each sample, we append LBP features just after HOG features. Besides, “our method” is short for the proposed CHDNet feature extraction method.
Figure 4Mean receiver operating characteristic of different feature extraction methods on tongue images dataset.
Comparison of the proposed method with different classifiers.
| Classifier | ACC | SEN | SPE | PPV | NPV |
|
|---|---|---|---|---|---|---|
| LDA | 84.90% | 91.09% | 50.42% | 91.16% | 51.10% | 91.08% |
| KNN | 89.55% | 93.10% | 69.71% | 94.56% | 65.18% | 93.78% |
| CART | 87.67% | 93.03% | 57.80% | 92.56% | 60.04% | 92.75% |
| GBDT | 87.91% | 93.18% | 58.60% | 92.65% | 62.72% | 92.87% |
| RF | 88.92% | 93.29% | 64.60% | 93.70% | 64.65% | 93.45% |
| LIBSVM | 91.27% | 94.76% | 72.04% | 95.00% | 72.90% | 94.83% |
| LIBLINEAR SVM | 91.14% | 94.22% | 75.40% | 95.59% | 74.56% | 94.83% |
Note. ACC = accuracy, SEN = sensitivity, SPC = specificity, PPV = positive predictive value, and NPV = negative predictive value.