| Literature DB >> 35977992 |
William Wallace1, Calvin Chan1, Swathikan Chidambaram1, Lydia Hanna1, Fahad Mujtaba Iqbal1,2, Amish Acharya1,2, Pasha Normahani1, Hutan Ashrafian2, Sheraz R Markar1,3,4, Viknesh Sounderajah5, Ara Darzi1,2.
Abstract
Digital and online symptom checkers are an increasingly adopted class of health technologies that enable patients to input their symptoms and biodata to produce a set of likely diagnoses and associated triage advice. However, concerns regarding the accuracy and safety of these symptom checkers have been raised. This systematic review evaluates the accuracy of symptom checkers in providing diagnoses and appropriate triage advice. MEDLINE and Web of Science were searched for studies that used either real or simulated patients to evaluate online or digital symptom checkers. The primary outcomes were the diagnostic and triage accuracy of the symptom checkers. The QUADAS-2 tool was used to assess study quality. Of the 177 studies retrieved, 10 studies met the inclusion criteria. Researchers evaluated the accuracy of symptom checkers using a variety of medical conditions, including ophthalmological conditions, inflammatory arthritides and HIV. A total of 50% of the studies recruited real patients, while the remainder used simulated cases. The diagnostic accuracy of the primary diagnosis was low across included studies (range: 19-37.9%) and varied between individual symptom checkers, despite consistent symptom data input. Triage accuracy (range: 48.8-90.1%) was typically higher than diagnostic accuracy. Overall, the diagnostic and triage accuracy of symptom checkers are variable and of low accuracy. Given the increasing push towards adopting this class of technologies across numerous health systems, this study demonstrates that reliance upon symptom checkers could pose significant patient safety hazards. Large-scale primary studies, based upon real-world data, are warranted to demonstrate the adequate performance of these technologies in a manner that is non-inferior to current best practices. Moreover, an urgent assessment of how these systems are regulated and implemented is required.Entities:
Year: 2022 PMID: 35977992 PMCID: PMC9385087 DOI: 10.1038/s41746-022-00667-w
Source DB: PubMed Journal: NPJ Digit Med ISSN: 2398-6352
Fig. 1Preferred reporting items for Systematic Reviews and Meta-Analysis (PRISMA) flow diagram showing the process of study selection for this systematic review of symptom checker diagnostic and triage accuracy.
Fig. 2Risk of bias summary using QUADAS-2 risk assessment tool[44].
Authors’ judgement regarding each domain of bias of each study synthesised on the accuracy of symptom checkers. Risk was categorised into one of three categories: low risk (+), unclear risk (?) and high risk (−). The table shows possible items to consider in future work involving symptom checkers to achieve a low risk of bias or applicability concerns.
Characteristics of the ten studies included in the systematic review on the diagnostic and triage accuracy of symptom checkers.
| First author (year) | Study design | Participants | Intervention | Symptom checker(s) used | Comparator | Outcome measures |
|---|---|---|---|---|---|---|
| Hageman (2015)[ | Prospective cross-sectional study | 86 hand clinic patients | Outpatients prospectively input data and symptoms into symptom checker to guess diagnosis | WebMD | Diagnosis from a hand surgeon | Diagnostic accuracy (top 3) |
| Semigran (2015)[ | Vignette cross-sectional study | 45 standardised vignettes: 15 emergency 15 non-emergency 15 self-care | Patient data and symptoms from various cases input into 23 symptom checkers | Ask MD, BetterMedicine. DocResponse, Doctor Diagnose, Drugs.com, EarlyDoc, Esagil, Family Doctor, FreeMD, HMS Family Health Guide, Healthline, Healthwise, Healthy Children, Isabel, iTriage, Mayo Clinic, MEDoctor, NHS Symptom Checkers, Steps2Care, Symcat, Symptify, Symptomate, WebMD | True diagnosis and appropriate triage advice | Diagnostic accuracy (primary diagnosis, top 3 and top 20 results) Triage accuracy (overall, emergency, non-emergency, self-care) |
| Powley (2016)[ | Prospective cross-sectional study | 34 inflammatory arthritis patients | Patients completed NHS and WebMD symptom checkers with their presenting symptoms | NHS Choices, WebMD | Diagnosis from secondary care | Diagnostic accuracy (primary diagnosis, top 5) |
| Berry (2019)[ | Retrospective cross-sectional study | ED records of 168 HIV/hepatitis C patients | Retrospective input of patient data into symptom checkers | Mayo Clinic, WebMD, Symptomate, Symcat, Isabel | Emergency department physician determined diagnosis | Diagnostic accuracy (primary diagnosis, top 3 and top 10) Triage accuracy |
| Nazario Arancibia (2019)[ | Prospective cross-sectional study | 214 low-priority ED patients | Patients interviewed using questions from symptom checker to produce differentials | Mediktor | Emergency department diagnosis | Diagnostic accuracy (primary diagnosis, top 3, top 5, and top 10) |
| Shen (2019)[ | Vignette cross-sectional study | 42 vignettes of ophthalmic conditions | Cases inputted into symptom checker to record results | WebMD | True diagnosis and appropriate triage advice | Diagnostic accuracy (primary diagnosis and top 3). Triage accuracy (emergency, non-emergency) |
| Gilbert (2020)[ | Vignette cross-sectional study | 200 primary care vignettes | Vignettes used to input into symptom checkers and to assess general practitioners (GPs) | ADA, Babylon, Buoy, K health, Mediktor, Symptomate, WebMD, YourMD | True diagnosis and GPs | Diagnostic accuracy (primary diagnosis, top 3 and top 5) Triage accuracy |
| Hill (2020)[ | Vignette cross-sectional study | 48 vignettes with ‘Australia-specific conditions’ | Vignettes summarised and entered in symptom checkers | 36 symptom checkers, 17 diagnostic only, 9 triage only advice. 10 that provide both. | True diagnosis and appropriate triage advice | Diagnostic accuracy (primary diagnosis, top 3 and top 10) Triage accuracy (emergency, non-emergency, self-care) |
| Yu (2020)[ | Retrospective cross-sectional study | 100 ED patient records | Retrospectively using patient records to input information into symptom checkers | Drugs.com FamilyDoctor | Given ED triage level | Triage accuracy |
| Yoshida (2021)[ | Vignette cross-sectional study | 27 vignettes of orofacial pain conditions | Vignettes of varied orofacial conditions input into symptom checkers | Esagil, FreeMD, Healthline, Isabel, Mayo Clinic, MEDoctor, Symcat, Symptify, Symptomate, WebMD, ADA | True diagnosis by resident in orofacial pain and oral medicine | Diagnostic accuracy (primary diagnosis, top 4) |
Overall and range of average diagnostic and triage accuracy of symptom checkers in each study.
| First author (year) | No. of symptom checkers | Overall average diagnostic accuracy (%) | Range of average diagnostic accuracy (%) | Average triage accuracy (%) | Range of average triage accuracy (%) |
|---|---|---|---|---|---|
| Hageman (2015)[ | 1 | 33 | n/a | ns | n/a |
| Semigran (2015)[ | 23 | 34 | 5–50 | 57 | 33–78 |
| Powley (2016)[ | 1 | 19 | n/a | ns | n/a |
| Berry (2019)[ | 5 | ns | 3–16.4 | 48.8 | ns |
| Nazario Arancibia (2019)[ | 1 | 37.9 | n/a | ns | n/a |
| Shen (2019)[ | 1 | 26 | n/a | 66.7 | n/a |
| Gilbert (2020)[ | 8 | 26.1 | 18–48 | 90.1 | 80–97.8 |
| Hill (2020)[ | 36 | 36 | 12–61 | 49 | 17–61 |
| Yu (2020)[ | 2 | ns | ns | 62 | 50–74 |
| Yoshida (2021)[ | 11 | 21.7 | 0–38.5 | ns | ns |
n/a not applicable as only one symptom checker was used, ns not stated.
Fig. 3Mean primary diagnostic accuracy of symptom checkers in each study.
Error bars signify the range of accuracy of different symptom checkers for the same patient/vignette population. An overall accuracy value was not given in Berry (2019).
Fig. 4Mean accuracy of triage information given by symptom checkers in each study.
Error bars signify the range of accuracy of different symptom checkers for the same patient/vignette population.
Primary diagnostic accuracy of WebMD in included studies.
| First author (year) | Participants | WebMD primary diagnostic accuracy (%) |
|---|---|---|
| Hageman (2015)[ | 86 hand clinic patients | 33 |
| Semigran (2015)[ | 45 standardised vignettes: | 36 |
| 15 emergency | ||
| 15 non-emergency | ||
| 15 self-care | ||
| Powley (2016)[ | 34 inflammatory arthritis patients | 19 |
| Berry (2019)[ | ED records of 168 HIV/hepatitis C patients | 3 (Hep C), 7.8 (HIV), 7.1 (both) |
| Shen (2019)[ | 42 vignettes of ophthalmic conditions | 26 |
| Gilbert (2020)[ | 200 primary care vignettes | 21 |
| Hill (2020)[ | 48 vignettes with ‘Australia-specific conditions’ | 53 |
| Yoshida (2021)[ | 27 vignettes of orofacial pain conditions | 30.77 |