| Literature DB >> 30425030 |
Ryan Rivas1, Niloofar Montazeri1, Nhat Xt Le1, Vagelis Hristidis1.
Abstract
BACKGROUND: An increasing number of doctor reviews are being generated by patients on the internet. These reviews address a diverse set of topics (features), including wait time, office staff, doctor's skills, and bedside manners. Most previous work on automatic analysis of Web-based customer reviews assumes that (1) product features are described unambiguously by a small number of keywords, for example, battery for phones and (2) the opinion for each feature has a positive or negative sentiment. However, in the domain of doctor reviews, this setting is too restrictive: a feature such as visit duration for doctor reviews may be expressed in many ways and does not necessarily have a positive or negative sentiment.Entities:
Keywords: patient reported outcome measures; patient satisfaction; quality indicators, health care; supervised machine learning
Mesh:
Year: 2018 PMID: 30425030 PMCID: PMC6256102 DOI: 10.2196/11141
Source DB: PubMed Journal: J Med Internet Res ISSN: 1438-8871 Impact factor: 5.428
Description of initial opinion classes. For each class, a sentence that does not mention the class is labeled c.
| Class | ||
| Appointment scheduling | Easy to schedule an appointment | Hard to schedule an appointment |
| Bedside manner | Friendly and caring | Rude and uncaring |
| Complementary medicine | Promotes complementary medicine | No promotion of complementary medicine |
| Cost | Inexpensive and billing is simple | Expensive and billing problems |
| Information sharing | Answers questions and good explanations | Does not answer questions and poor explanations |
| Joint decision making | Treatment plan accounts for patient opinions | Treatment plan made without patient input |
| Medical skills | Effective treatment and correct diagnoses | Ineffective treatment and misdiagnoses conditions |
| Psychological support | Addresses stress and anxiety | Does not address stress and anxiety |
| Self-management | Encourages active management of care | Does not encourage self-management of care |
| Staff | Staff is friendly and helpful | Staff is rude and unhelpful |
| Technology | Uses email, Web-based appointments, and electronic health records | Does not use email and Web-based appointments |
| Visit time | Spends substantial time with patients | Spends very little time with patients |
| Wait time | Short time spent waiting to see the doctor | Long time spent waiting to see the doctor |
Figure 1Screenshot of WebAnno’s annotation interface with an annotated review.
Frequency of each class label in the doctor review dataset.
| Class | Frequency of | Frequency of | Frequency of |
| 51 | 84 | 5750 | |
| 569 | 341 | 4975 | |
| 25 | 261 | 5599 | |
| 316 | 136 | 5433 | |
| 481 | 232 | 5172 | |
| 262 | 368 | 5255 | |
| 143 | 79 | 5663 | |
| 48 | 199 | 5638 |
Figure 2A dependency tree for the sentence "There are never long wait times".
Figure 3Dependency tree for the sentence "I arrived to my appointment on time and waited in his waiting room for over an hour".
Weighted accuracy of classifiers on doctor review dataset.
| Classifier | Average (%) | ||||||||
| CNNa | 42.06 | 56.69 | 42.75 | 51.45 | 47.81 | 61.42 | 55.38 | 60.93 | 52.31 |
| CNN-Wb | 49.89 | 44.30 | 53.53 | 49.71 | 64.04 | 54.29 | 63.51 | 54.87 | |
| D2V-NNd | 38.83 | 45.16 | 38.00 | 42.25 | 41.44 | 42.19 | 41.04 | 43.64 | 41.57 |
| Matsumoto | 45.76 | 59.63 | 45.89 | 53.40 | 49.89 | 57.24 | 55.83 | ||
| RFe | 40.78 | 42.00 | 34.76 | 37.29 | 41.62 | 52.88 | 45.65 | 51.66 | 43.33 |
| SVMf | 33.33 | 35.77 | 33.33 | 33.33 | 33.33 | 48.94 | 33.33 | 48.07 | 37.43 |
| DTCg | 51.72 | 50.48 | 41.27 | 47.23 | 38.49 | 54.31 | 65.91 | 51.29 | |
| DTCRF | 46.64 | 39.19 | 47.29 | 40.20 | 56.15 | 60.57 | 58.05 | 50.26 | |
| DTCCNN-W | 53.89 | 59.37 | 61.43 | 56.63 | 67.67 |
aCNN: Convolutional Neural Network.
bCNN-W: Convolutional Neural Network with Word2Vec.
cThe highest value for each c is italicized for emphasis.
dD2V-NN: Doc2Vec Nearest Neighbor.
eRF: Random Forests.
fSVM: Support Vector Machine.
gDTC: dependency tree classifier.
Per-label accuracy of top 3 classifiers on doctor review dataset for each c and c.
| Label and classifier | |||||||||
| CNN-Wa | 31.37% | 57.22% | 0.00% | 47.62% | 40.54% | 60.69% | 45.07% | 40.85% | |
| Matsumoto | 13.73% | 57.04% | 48.57% | 41.16% | 59.16% | 47.89% | |||
| DTCcCNN-W | 39.44% | ||||||||
| CNN-W | 19.05% | 27.35% | 34.48% | 15.44% | 13.36% | 35.42% | 18.99% | 50.75% | |
| Matsumoto | 23.81% | 27.65% | 35.00% | 13.24% | 12.93% | 20.25% | |||
| DTCCNN-W | 27.52% | 35.68% | |||||||
aCNN-W: Convolutional Neural Network with Word2Vec.
bFor each c, the highest value for both c and c are italicized for emphasis.
cDTC: dependency tree classifier.
Ratio of sentences classified by the top 3 classifiers as c or c that were classified correctly.
| Label and classifier | |||||||||
| CNN-Wa | 34.78% | 0.00% | 62.50% | 50.26% | 66.81% | 57.14% | 65.91% | ||
| Matsumoto | 43.40% | ||||||||
| DTCcCNN-W | 16.04% | 41.66% | 10.00% | 20.69% | 22.58% | 43.59% | 23.73% | 21.52% | |
| CNN-W | 40.00% | 50.56% | 22.83% | 41.27% | 29.41% | 59.06% | |||
| Matsumoto | 34.18% | 25.64% | |||||||
| DTCCNN-W | 10.98% | 13.50% | 28.57% | 13.38% | 14.25% | 22.90% | 14.29% | 29.96% | |
aCNN-W: Convolutional Neural Network with Word2Vec.
bFor each c, the highest value for both c and c are italicized for emphasis.
cDTC: dependency tree classifier.