| Literature DB >> 30021708 |
Yong Shi1,2,3,4, Peijia Li5,6, Xiaodan Yu7, Huadong Wang2,3, Lingfeng Niu1,2,3.
Abstract
BACKGROUND: Doctor's performance evaluation is an important task in mobile health (mHealth), which aims to evaluate the overall quality of online diagnosis and patient outcomes so that customer satisfaction and loyalty can be attained. However, most patients tend not to rate doctors' performance, therefore, it is imperative to develop a model to make doctor's performance evaluation automatic. When evaluating doctors' performance, we rate it into a score label that is as close as possible to the true one.Entities:
Keywords: mHealth; ordinal partitioning; ordinal regression; performance evaluation; support vector machines
Mesh:
Year: 2018 PMID: 30021708 PMCID: PMC6070724 DOI: 10.2196/jmir.9300
Source DB: PubMed Journal: J Med Internet Res ISSN: 1438-8871 Impact factor: 5.428
Figure 1The general workflow of the ordinal regression for doctor performance evaluation (OR-DPE) model.
The details about the medical dictionaries. “1≤terms≤3” means the number of terms having a character length less than 3 but greater than 1.
| Number of phrases | Dictionary Name | |
| Illness and Symptom Dictionary (N=49,758) | Medicine Dictionary (N=24,975) | |
| 1≤terms≤3, n (%) | 32840 (66.00) | 3746 (15.00) |
| 4≤terms≤6, n (%) | 16918 (34.00) | 14486 (58.00) |
| terms≥7, n (%) | 0 (0) | 6743 (27.00) |
| Representative examples | 神经衰弱症 (Neurosis), 高血压 (HTN), 天花 (smallpox) | 帕罗西汀 (Paroxetine), 盐酸环苯扎林 (Flexeril) |
F1-F8 represent the customized medical features, while F9 and F10 are the text features.
| Feature | Description |
| F1 | The number of symptom names in doctors’ answers |
| F2 | The number of illness names in doctors’ answers |
| F3 | The number of medicine names in doctors’ answers |
| F4 | The number of patients’ questions |
| F5 | The number of doctors’ answers |
| F6 | The response time for the patient’s first question |
| F7 | The number of Chinese characters in patients’ questions |
| F8 | The number of Chinese characters in doctors’ answers |
| F9 | Unigrams |
| F10 | Bigrams |
Figure 2The demo that shows how a combined support vector machine and ordinal partitioning scheme model (SVMOP) works on ordinal data.
Figure 3An example of consultation letters on Platform X. The left subfigure is the real consultation on Platform X by mobile software applications but without sensitive information such as doctors’ photos. The right one is the version in English.
Performances of various models having multiple feature sets (T, T+C, T+C+B) are shown in this table.
| Method | Text (T) | Text and Customized (T+C) | Text, Customized, and Booster (T+C+B) | |||||||
| MAEa | MSEb | PAccc (%) | MAE | MSE | PAcc (%) | MAE | MSE | PAcc (%) | ||
| SVCd | 0.7925 | 1.7613 | 53.32 | 0.6726 | 1.3759 | 57.32 | 0.6212e | 1.1981e | 59.05e | |
| SVRf | 0.8023 | 1.3302 | 49.74 | 0.7050 | 1.1106 | 54.24 | 0.6906e | 1.0332e | 56.37e | |
| LRg | 0.7716 | 1.6883 | 53.86 | 0.6359 | 1.2606 | 57.77 | 0.5978e | 1.1310e | 59.50e | |
| SVORh | 0.8086 | 1.3742 | 49.58 | 0.7170 | 1.1167 | 54.09 | 0.6665e | 1.0143e | 57.20e | |
| RedSVMi | 0.8046 | 1.3715 | 50.11 | 0.7168 | 1.1127 | 54.00 | 0.6718e | 1.0236e | 57.21e | |
| SVMOPj | 0.7054k | 1.2706k | 54.11k | 0.6130k | 1.0108k | 57.92k | 0.5864e,k | 0.9605e,k | 59.65e,k | |
aMAE: mean absolute error.
bMSE: mean standard error.
cPAcc: pairwise accuracy.
dSVC: support vector classification.
eBest “one of” feature sets.
fSVR: support vector regression.
gLR: logistic regression.
hSVOR: support vector ordinal regression.
iRedSVM: reduction support vector machine.
jSVMOP: a combined support vector machine and ordinal partitioning scheme model.
kBest performance for each metric.
Figure 4Mean absolute error (MAE), mean square error (MSE), and pairwise accuracy (PAcc) varying from different models and different feature sets. LR: logistic regression; RedSVM: reduction support vector machine; SVC: support vector classification; SVMOP: a combined support vector machine and ordinal partitioning scheme model; SVOR: support vector ordinal regression; T: text features; T+C: text and customized features; T+C+B: text, customized, and boosted features.
Figure 5The different performances with different parameters in training process with the text, customized, and boosted feature set (T+C+B). LR: logistic regression; MAE: mean absolute error; MSE: mean square error; PAcc: pairwise accuracy; RedSVM: reduction support vector machine; SVC: support vector classification; SVMOP: a combined support vector machine and ordinal partitioning scheme model; SVOR: support vector ordinal regression; SVR: support vector regression.
Figure 6The confusion matrices of different models with the text, customized, and boosted feature set (T+C+B). RedSVM: reduction support vector machine; SVC: support vector classification; SVMOP: a combined support vector machine and ordinal partitioning scheme model; SVOR: support vector ordinal regression; SVR: support vector regression.
Benchmark datasets. “#ins” is the number of instances. “#fea” is the number of features. “#class” is the number of classes.
| Datasets | #ins | #fea | #class |
| housing-5 | 10120 | 14 | 5 |
| machine-5 | 4180 | 7 | 5 |
| abalone-5 | 83540 | 11 | 5 |
| housing-10 | 10120 | 14 | 10 |
| machine-10 | 4180 | 7 | 10 |
| abalone-10 | 83540 | 11 | 10 |
The mean absolute error (MAE), mean standard error (MSE), and pairwise accuracy (PAcc) performances of different models on benchmark datasets. The best result is indicated by a footnote.
| Datasets | SVCa | SVRb | LRc | SVORd | RedSVMe | SVMOPf | |
| housing-5 | 0.517 | 0.454 | 0.435 | 0.398 | 0.403 | 0.366g | |
| machine-5 | 0.606 | 0.550 | 0.451 | 0.390 | 0.424 | 0.369g | |
| abalone-5 | 0.798 | 0.712 | 0.700 | 0.683 | 0.675 | 0.648g | |
| housing-10 | 1.513 | 0.962 | 0.999 | 0.859 | 0.848 | 0.757g | |
| machine-10 | 1.425 | 1.151 | 0.986 | 0.935 | 0.927 | 0.841g | |
| abalone-10 | 1.959 | 1.451 | 1.557 | 1.435 | 1.434 | 1.391g | |
| housing-5 | 0.665 | 0.545 | 0.612 | 0.494 | 0.524 | 0.446g | |
| machine-5 | 0.994 | 0.634 | 0.648 | 0.469 | 0.505 | 0.429g | |
| abalone-5 | 1.450 | 0.992 | 1.244 | 1.042 | 0.991 | 0.962g | |
| Housing-10 | 4.564 | 1.858 | 2.560 | 1.694 | 1.642 | 1.453g | |
| machine-10 | 3.998 | 2.487 | 2.277 | 1.786 | 1.720 | 1.547g | |
| abalone-10 | 7.222 | 3.703 | 5.091 | 3.586 | 3.783 | 3.635g | |
| housing-5 | 0.614 | 0.638 | 0.658 | 0.663 | 0.659 | 0.676g | |
| machine-5 | 0.602 | 0.604 | 0.652 | 0.666 | 0.655 | 0.680g | |
| abalone-5 | 0.547 | 0.553 | 0.584 | 0.584 | 0.577 | 0.589g | |
| Housing-10 | 0.552 | 0.623 | 0.609 | 0.635 | 0.637 | 0.642g | |
| machine-10 | 0.488 | 0.562 | 0.597 | 0.601 | 0.599 | 0.612g | |
| abalone-10 | 0.514 | 0.568 | 0.566 | 0.565 | 0.568 | 0.569g | |
aSVC: support vector classification.
bSVR: support vector regression.
cLR: logistic regression.
dSVOR: support vector ordinal regression.
eRedSVM: reduction support vector machine.
fSVMOP: a combined support vector machine and ordinal partitioning scheme model.
gBest result.