| Literature DB >> 23226331 |
Razan Paul1, Tudor Groza, Jane Hunter, Andreas Zankl.
Abstract
A lack of mature domain knowledge and well established guidelines makes the medical diagnosis of skeletal dysplasias (a group of rare genetic disorders) a very complex process. Machine learning techniques can facilitate objective interpretation of medical observations for the purposes of decision support. However, building decision support models using such techniques is highly problematic in the context of rare genetic disorders, because it depends on access to mature domain knowledge. This paper describes an approach for developing a decision support model in medical domains that are underpinned by relatively sparse knowledge bases. We propose a solution that combines association rule mining with the Dempster-Shafer theory (DST) to compute probabilistic associations between sets of clinical features and disorders, which can then serve as support for medical decision making (e.g., diagnosis). We show, via experimental results, that our approach is able to provide meaningful outcomes even on small datasets with sparse distributions, in addition to outperforming other Machine Learning techniques and behaving slightly better than an initial diagnosis by a clinician.Entities:
Mesh:
Year: 2012 PMID: 23226331 PMCID: PMC3511538 DOI: 10.1371/journal.pone.0050614
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Figure 1Relative distribution of dysplasia diagnoses according to different ranges of number of cases.
More than 70% of the bone dysplasias present in the ESDN dataset have a very small number of cases (up to 5), while those that are well represented (i.e., over 50 cases) represent a mere fraction of the total number –4%.
Coverage of clinical and radiographic features in the ESDN dataset.
| Clinical Feature | No. cases | Total diagnoses | Coverage (%) |
| Cystic hygroma | 2 | 4 | 50 |
| Subglottic stenosis | 1 | 2 | 50 |
| Cyanosis | 1 | 2 | 50 |
| Hypopigmentation of the skin | 1 | 101 | 0.99 |
| Immunodeficiency | 1 | 101 | 0.99 |
| Clumsiness | 1 | 101 | 0.99 |
The table presents the top 3 and bottom 3 coverages. Coverage is computed by dividing the number of cases that contain the clinical feature by the total number of diagnoses denoting the bone dysplasias assigned to the cases. For example, Cystic hygroma appears in 2 of the total 4 cases diagnosed with Achondrogenesis type 1A.
Characteristics of the skeletal dysplasias with more than 20 cases.
| Diagnosis | Symbol | No. cases | Total features | Min | Max | Average | Max coverage | Largest common set coverage |
| Hypochondroplasia |
| 22 | 69 | 1 | 15 | 5 (7.24%) | 54.54% | 27% (3) |
| SEDC |
| 75 | 151 | 1 | 17 | 4.65 (3%) | 40% | 26% (2) |
| Pseudoachondroplasia |
| 33 | 72 | 1 | 12 | 4.15 (5.76%) | 57.77% | 12% (3) |
| Cartilage-hair-hypoplasia |
| 28 | 80 | 1 | 11 | 4.89 (6.11%) | 46.42% | 10% (4) |
| MED (AD) |
| 101 | 128 | 1 | 20 | 3 (2.34%) | 28.71% | 19% (3) |
| rMED |
| 24 | 59 | 1 | 13 | 3.87 (6.55%) | 25% | 25% (2) |
The set of dysplasias used within our experiments follow, in principle, the general characteristics of the ESDN dataset. The average maximum coverage of phenotypes is around 43%, while the average largest common set (i.e., the set of phenotypes common to all cases diagnosed with a particular disorder) is around 20%.
Experimental results: Accuracy per cross-validation per fold.
|
|
|
|
|
| Interval | Mean (%) | |
| Accuracy (%) | 41.5 | 46 | 49.05 | 52.97 | 47.16 | [41.5, 52.94] | 47.43 |
| Averageprecision (%) | 31.08 | 51.63 | 51.61 | 41.69 | 20.71 | [20.71, 51.63] | 39.34 |
| Averagerecall (%) | 28.4 | 38.67 | 38.31 | 42.11 | 26.95 | [26.95, 42.11] | 34.89 |
Experimental results: Overall comparative accuracy across all considered approaches.
| Our approach | Naive Bayes | SVM | Decision trees | Random forests | k-NN (K = 3) | |||||||
| P (%) | R (%) | P (%) | R (%) | P (%) | R (%) | P (%) | R (%) | P (%) | R (%) | P (%) | R (%) | |
|
| 6.67 | 5 | 20 | 5 | 11.66 | 10 | 13.16 | 15 | 15.54 | 15 | 10 | 5 |
|
| 47.71 | 70.12 | 53.66 | 34.72 | 38.12 | 40.46 | 29.78 | 32.98 | 36.98 | 33.08 | 31.92 | 38.34 |
|
| 42.41 | 27.78 | 41.66 | 9.16 | 36.42 | 25.26 | 25.18 | 26.94 | 25.8 | 24.98 | 5.84 | 11.66 |
|
| 50.67 | 32 | 20 | 4 | 33.08 | 32 | 44 | 24 | 21.42 | 20 | 20 | 4 |
|
| 58.47 | 59.43 | 41.94 | 97.64 | 47.07 | 64.14 | 52.88 | 45.52 | 50.64 | 59.2 | 45.04 | 62.52 |
|
| 30 | 15 | 0 | 0 | 40 | 15 | 21.68 | 31.66 | 10 | 10 | 0 | 0 |
| Average recall (%) | 34.89 | 25.08 | 31.14 | 29.35 | 27.55 | 20.25 | ||||||
| Average prec. (%) | 39.34 | 29.54 | 34.38 | 31.11 | 26.8 | 18.8 | ||||||
| Accuracy rate (%) | 47.43 | 44.12 | 41.08 | 33.83 | 36.89 | 34.23 | ||||||
Our solution outperforms the five Machine Learning approaches we have considered within our experiments: around 4% more accuracy than Naive Bayes and around 6% more accuracy than SVM. Although Naive Bayes has performed very well, its results are boosted by overfitting the classes that had more data (e.g., ) at the expense of others, such as for which it achieved 0 precision and recall. Unlike Naive Bayes, our approach has performed fairly uniform and consistent across all classes.