Literature DB >> 26769236

Crowdsourcing Diagnosis for Patients With Undiagnosed Illnesses: An Evaluation of CrowdMed.

Ashley N D Meyer1, Christopher A Longhurst, Hardeep Singh.   

Abstract

BACKGROUND: Despite visits to multiple physicians, many patients remain undiagnosed. A new online program, CrowdMed, aims to leverage the "wisdom of the crowd" by giving patients an opportunity to submit their cases and interact with case solvers to obtain diagnostic possibilities.
OBJECTIVE: To describe CrowdMed and provide an independent assessment of its impact.
METHODS: Patients submit their cases online to CrowdMed and case solvers sign up to help diagnose patients. Case solvers attempt to solve patients' diagnostic dilemmas and often have an interactive online discussion with patients, including an exchange of additional diagnostic details. At the end, patients receive detailed reports containing diagnostic suggestions to discuss with their physicians and fill out surveys about their outcomes. We independently analyzed data collected from cases between May 2013 and April 2015 to determine patient and case solver characteristics and case outcomes.
RESULTS: During the study period, 397 cases were completed. These patients previously visited a median of 5 physicians, incurred a median of US $10,000 in medical expenses, spent a median of 50 hours researching their illnesses online, and had symptoms for a median of 2.6 years. During this period, 357 active case solvers participated, of which 37.9% (132/348) were male and 58.3% (208/357) worked or studied in the medical industry. About half (50.9%, 202/397) of patients were likely to recommend CrowdMed to a friend, 59.6% (233/391) reported that the process gave insights that led them closer to the correct diagnoses, 57% (52/92) reported estimated decreases in medical expenses, and 38% (29/77) reported estimated improvement in school or work productivity.
CONCLUSIONS: Some patients with undiagnosed illnesses reported receiving helpful guidance from crowdsourcing their diagnoses during their difficult diagnostic journeys. However, further development and use of crowdsourcing methods to facilitate diagnosis requires long-term evaluation as well as validation to account for patients' ultimate correct diagnoses.

Entities:  

Keywords:  World Wide Web; crowdsourcing; diagnosis; diagnostic errors; patient safety

Mesh:

Year:  2016        PMID: 26769236      PMCID: PMC4731679          DOI: 10.2196/jmir.4887

Source DB:  PubMed          Journal:  J Med Internet Res        ISSN: 1438-8871            Impact factor:   5.428


Introduction

Errors of clinical diagnosis affect at least 5% of US adults every year and approximately half of these errors could result in serious harm to the patients [1]. To address the extent and severity of this problem, both systems and cognitive solutions have been proposed. However, only a few of these have been tested and only a fraction of those tested have been shown to improve diagnostic outcomes [2-4]. Patients with difficult-to-diagnose conditions often seek care from several physicians and institutions before obtaining a diagnosis. One intervention that could benefit patients is the use of second opinions [5-7], and this has been shown to catch previously missed diagnoses, at least in the realms of radiology and pathology [6]. Several formal programs currently exist to provide second opinions to patients. [7] For example, in the NIH Undiagnosed Diseases Network based at several centers across the US [8], medical experts diagnose undiagnosed individuals or those with rare diseases. The program, however, has strict eligibility requirements for patients and requires a clinician referral. Additional programs include Best Doctors’ second-opinion program that is open to employee beneficiaries only and Cleveland Clinic’s MyConsult program [5,9], both of which involve comprehensive review of patients’ medical records, but no dynamic interactions with the patients. A recently developed software platform, CrowdMed [10], aims to overcome some limitations of the aforementioned programs, namely; strict eligibility requirements, needed referrals, and limited interaction with patients; by leveraging the “wisdom of the crowd” or crowdsourcing to help undiagnosed or misdiagnosed patients. Crowdsourcing is a “participative online activity” in which a group of individuals of varying knowledge, heterogeneity, and number comes together to solve a problem [11]. It has been used for a variety of problems in different fields ranging from simple text translation to more complicated tasks, such as solving the BP oil spill disaster in the Gulf of Mexico [12]. In medicine, it has been utilized for health and medical research, such as estimating flu prevalence [13]; for informatics solutions, including establishing problem-related medication pairs [14], and for examining specific diseases through image analysis. In the latter situation, crowdsourcing has been used to inspect blood samples to determine the presence or absence of malarial infections [15-17] and to categorize colorectal polyps [18,19] or diabetic retinopathy [20]. However, until now crowdsourcing had not been used to come up with a diagnosis from all possible diagnoses a patient might have. Of note, this platform allows laypersons without health care training or experience to participate. Although patients have been “googling for a diagnosis” for more than a decade and even using online symptom checkers [21,22], this is the first description of a crowd of people working together online towards a more accurate diagnosis. We conducted an independent evaluation of this untested approach to determine whether this could be beneficial to patient care.

Methods

A Description of CrowdMed

For a small fee, the CrowdMed website allows undiagnosed patients to submit their clinical information and obtain potential diagnoses expeditiously. Patients anonymously answer a comprehensive set of medical questions and upload relevant test results and images related to their cases (Figure 1).
Figure 1

Screenshot of case submission.

Patients also decide how long they want their cases open and whether they wish to compensate the case solvers. Anyone (including nonmedical persons) can sign up to be a case solver and select cases they think they can help solve (Figure 2).
Figure 2

Screenshot of case selection for solvers (names are fictitious).

While the cases are open, patients and case solvers can discuss details online about potential diagnoses, further work-up that should be done, and newly obtained test results and/or appointments completed with the patients’ physicians. Thus, case details can unfold online while the case is still open. All diagnostic suggestions and all case discussions are available to all case solvers as they are suggested and discussed throughout the open period. This enables the entire group of case solvers to work in concert to solve each case. When a patient’s case is closed, the patient receives a detailed report containing the entire list of diagnostic suggestions made by the case solvers and suggested next steps, so that they can discuss them with their physicians. Diagnoses are ranked in decreasing order of “relative popularity.” The relative popularity of diagnoses is determined by case solvers’ “bets” on each diagnosis in terms of their beliefs that the diagnosis is the most specific, accurate, root cause of the symptoms presented. CrowdMed takes these bets and assigns points to each diagnosis using a prediction market algorithm, thereby determining the “relative popularity” of each diagnosis suggested. Finally, patients are provided with case solvers’ reasoning for choosing particular diagnoses. Patients choose which case solver(s) to compensate based on whose answers they found helpful. If the patient decides to reward multiple solvers, they also decide how to divvy up the compensation. Afterward, patients are invited to fill out surveys about their outcomes. Screenshot of case submission. Screenshot of case selection for solvers (names are fictitious).

Independent Evaluation

We independently analyzed all CrowdMed data collected from May 2013 to April 2015. Specifically, we analyzed data on patients’ demographic and case characteristics; case solvers’ demographic and performance characteristics; and preliminary case outcomes. Outcomes included whether patients would recommend CrowdMed, if the program provided insights leading them closer to correct diagnoses, and estimated improvements in patients’ productivity and medical expenses. Data were summarized using descriptive statistics and independent samples t tests using IBM SPSS Statistics 22.

Results

Patients and Cases

During the study period, 397 cases were completed (350 from the United States). Patients’ self-reported mean (SD) age was 47.8 (18.8) years (age range 2-90) and 182 were males (45.8%). Before case submission, patients reported visiting a median of 5 physicians (interquartile range [IQR] 3-10; range 0-99), incurred a median of US $10,000 in medical expenses (IQR US $2500-US $50,000; range US $0-US $5,000,000) including payments by both patients and payers, spent a median of 50 hours (IQR 15-150; range 0-12,000) researching their illnesses online, and had symptoms for a median of 2.6 years (IQR 1.1-6.9; range 0.0-70.6). Online case activity lasted a median of 60 days (IQR 30-90; range 2-150) and case solvers were offered a median of US $100 in compensation (IQR US $0-US $200; range US $0-US $2700) for diagnostic suggestions. A total of 59.7% (237/397) of the cases were compensated with a median compensation of US $200 (IQR US $100-US $300; range US $15-US $2700).

Case Solvers

During the study period, CrowdMed had 357 active case solvers; of which 37.9% (132/348) were male, 76.7% (264/344) were from the US, and 58.3% (208/357) worked or studied in the medical industry; including 36 physicians and 56 medical students. Mean (SD) age was 39.6 (13.8) years (range 17-77 years). Solvers participated in a median of 3 cases (IQR 1.0-12.8; range 0-415), earned a median of US $0 (IQR US $0-US $1.18; range US $0-US $3952) and a mean (SD) of US $93.97 (US $364.72; the majority earned US $0). Median solver rating was 3 (out of 10; IQR 3-6; range 1-10) and significantly higher (P=.006) for medical industry-based solvers (mean [SD] 4.8 [2.5]; range 1-10) than for others (mean [SD] 4.1 [2.2]; range 1-10).

Outcomes

At completion, 50.9% (202/397) of patients were likely to recommend CrowdMed to a friend, 59.6% (233/391) reported that the process gave insights leading them closer to correct diagnoses, 57% (52/92) reported estimated decreases in medical expenses, and 38% (29/77) reported estimated improvements in school or work productivity (Table 1).
Table 1

Case outcomes as assessed in a postcase survey.

Case outcomes n (%)
On a scale of 1-5, How likely are you to recommend CrowdMed to a friend (with 5 being most likely)? (391/397 surveyed answered; 98.5% response rate)
139 (10.0)
243 (11.0)
3107 (27.4)
476 (19.4)
5126 (32.2)
Did CrowdMed Medical Detective community provide insights that lead you closer to a correct diagnosis or cure? (391/397 surveyed answered; 98.5% response rate)
No158 (40.4)
Yes233 (59.6)
How much do you estimate that your CrowdMed results will reduce the cost of your medical case going forward? (92/147 surveyed answered; 62.6% response rate) a
1-20%25 (27.2)
21-50%15 (16.3)
51-80%10 (10.9)
>80%2 (2.2)
Not at all40 (43.5)
How much lost work or school productivity do you estimate that your CrowdMed results will help you regain going forward? (77/147 surveyed answered; 52.4% response rate) a
1-20%12 (15.6)
21-50%8 (10.4)
51-80%7 (9.1)
81-99%1 (1.3)
All1 (1.3)
None48 (62.3)

aThese questions were added to the postcase survey later.

Case outcomes as assessed in a postcase survey. aThese questions were added to the postcase survey later. Patients reporting helpful insights from CrowdMed saw fewer doctors (mean [SD] 7.2 [7.3]; range 0-99) before participating than those who did not report receiving helpful insights (mean [SD] 9.2 [10.7]; range 0-50), P=.047. The 14 most common diagnoses suggested as the most popular diagnosis for a case are presented in Table 2.
Table 2

The 14 most common diagnoses suggested as the most popular diagnosis across 397 cases.

Diagnosisn (%)
Lyme disease8 (2.0)
Dysautonomia7 (1.8)
Chronic fatigue syndrome6 (1.5)
Irritable bowel syndrome6 (1.5)
Mast cell activation disorder6 (1.5)
Postural orthostatic tachycardia syndrome5 (1.3)
Ehlers-Danlos syndrome4 (1.0)
Sjögren’s syndrome4 (1.0)
Abdominal cutaneous nerve entrapment syndrome3 (0.8)
Gastroesophageal reflux disease3 (0.8)
Hypothyroidism3 (0.8)
Multiple sclerosis3 (0.8)
Myasthenia gravis3 (0.8)
The 14 most common diagnoses suggested as the most popular diagnosis across 397 cases. In addition, some patients informally reported to CrowdMed that the program helped them find diagnoses that their physicians previously were unable to determine, including Sjögren’s syndrome and chorda tympani dysfunction.

Discussion

Main Findings

Our independent evaluation suggests that at least some patients with undiagnosed illnesses reported receiving helpful guidance from crowdsourcing their diagnoses during their difficult diagnostic journeys. Several of the conditions most commonly suggested by case solvers are conditions well known to represent diagnostic challenges. The crowdsourcing strategy enabled dynamic interaction between patients and case solvers as more case details unfolded over time. Novel approaches are needed to help patients who experience difficulties in obtaining a correct and timely diagnosis. In that regard, advantages of using “wisdom of the crowd” could include low cost, increased program accessibility for patients, and relatively quick opinions. Although the data we obtained were useful for understanding this program, there were several limitations of our study. The postparticipation survey was rather limited in scope as it was designed for business purposes and not for research. In addition, there was no way to verify patient-reported data and some patient-reported data might be outside of realistic boundaries (eg, 1 patient reported spending 12,000 hours researching illnesses online). Furthermore, downstream outcomes of patients were not systematically collected, so it is not known what their eventual diagnoses were or if the program identified them accurately. Further development and use of crowdsourcing methods to facilitate diagnosis requires long-term evaluation as well as validation to account for patients’ ultimate correct diagnoses. Although crowdsourcing appears to have potential, it is important to identify factors that lead to successful crowdsourcing to improve the process and help improve patient care. Multidisciplinary research is needed to gain both technical and nontechnical insights into how this can be done. For example, previous researchers have identified the importance of both finding crowd members with the appropriate skills to the relevant problem and providing adequate motivation to the crowd for the successful use of crowdsourcing for problem solving [23]. Finally, the potential legal ramifications of giving individuals without medical degrees (who make up a substantial portion of the case solvers) the ability to render diagnostic opinions would need to be considered [24].

Conclusions

In conclusion, our independent evaluation suggests that some patients with undiagnosed illnesses report receiving helpful guidance from crowdsourcing their diagnosis. Further development and use of crowdsourcing methods to facilitate diagnosis require multidisciplinary research and long-term evaluation that includes validation to account for patients’ ultimate correct diagnoses.
  16 in total

1.  Distributed human intelligence for colonic polyp classification in computer-aided detection for CT colonography.

Authors:  Tan B Nguyen; Shijun Wang; Vishal Anugu; Natalie Rose; Matthew McKenna; Nicholas Petrick; Joseph E Burns; Ronald M Summers
Journal:  Radiology       Date:  2012-01-24       Impact factor: 11.105

2.  Googling for a diagnosis--use of Google as a diagnostic aid: internet based study.

Authors:  Hangwi Tang; Jennifer Hwee Kwoon Ng
Journal:  BMJ       Date:  2006-11-10

Review 3.  Crowdsourcing--harnessing the masses to advance health and medicine, a systematic review.

Authors:  Benjamin L Ranard; Yoonhee P Ha; Zachary F Meisel; David A Asch; Shawndra S Hill; Lance B Becker; Anne K Seymour; Raina M Merchant
Journal:  J Gen Intern Med       Date:  2013-07-11       Impact factor: 5.128

Review 4.  System-related interventions to reduce diagnostic errors: a narrative review.

Authors:  Hardeep Singh; Mark L Graber; Stephanie M Kissam; Asta V Sorensen; Nancy F Lenfestey; Elizabeth M Tant; Kerm Henriksen; Kenneth A LaBresh
Journal:  BMJ Qual Saf       Date:  2011-11-30       Impact factor: 7.035

Review 5.  Patient-initiated second opinions: systematic review of characteristics and impact on diagnosis, treatment, and satisfaction.

Authors:  Velma L Payne; Hardeep Singh; Ashley N D Meyer; Lewis Levy; David Harrison; Mark L Graber
Journal:  Mayo Clin Proc       Date:  2014-05       Impact factor: 7.616

6.  Crowd-sourced BioGames: managing the big data problem for next-generation lab-on-a-chip platforms.

Authors:  Sam Mavandadi; Stoyan Dimitrov; Steve Feng; Frank Yu; Richard Yu; Uzair Sikora; Aydogan Ozcan
Journal:  Lab Chip       Date:  2012-10-21       Impact factor: 6.799

7.  Distributed medical image analysis and diagnosis through crowd-sourced games: a malaria case study.

Authors:  Sam Mavandadi; Stoyan Dimitrov; Steve Feng; Frank Yu; Uzair Sikora; Oguzhan Yaglidere; Swati Padmanabhan; Karin Nielsen; Aydogan Ozcan
Journal:  PLoS One       Date:  2012-05-11       Impact factor: 3.240

8.  Rapid grading of fundus photographs for diabetic retinopathy using crowdsourcing.

Authors:  Christopher J Brady; Andrea C Villanti; Jennifer L Pearson; Thomas R Kirchner; Omesh P Gupta; Chirag P Shah
Journal:  J Med Internet Res       Date:  2014-10-30       Impact factor: 5.428

9.  Crowdsourcing malaria parasite quantification: an online game for analyzing images of infected thick blood smears.

Authors:  Miguel Angel Luengo-Oroz; Asier Arranz; John Frean
Journal:  J Med Internet Res       Date:  2012-11-29       Impact factor: 5.428

10.  The frequency of diagnostic errors in outpatient care: estimations from three large observational studies involving US adult populations.

Authors:  Hardeep Singh; Ashley N D Meyer; Eric J Thomas
Journal:  BMJ Qual Saf       Date:  2014-04-17       Impact factor: 7.035

View more
  16 in total

1.  Understanding Patient Questions about their Medical Records in an Online Health Forum: Opportunity for Patient Portal Design.

Authors:  Tera L Reynolds; Nida Ali; Emma McGregor; Trish O'Brien; Christopher Longhurst; Andrew L Rosenberg; Scott E Rudkin; Kai Zheng
Journal:  AMIA Annu Symp Proc       Date:  2018-04-16

2.  [Computer-assisted diagnosis of rare diseases].

Authors:  T Müller; A Jerrentrup; J R Schäfer
Journal:  Internist (Berl)       Date:  2018-04       Impact factor: 0.743

3.  A systematic review of natural language processing and text mining of symptoms from electronic patient-authored text data.

Authors:  Caitlin Dreisbach; Theresa A Koleck; Philip E Bourne; Suzanne Bakken
Journal:  Int J Med Inform       Date:  2019-02-20       Impact factor: 4.046

4.  Rare disease patient matchmaking: development and outcomes of an internet case-finding strategy in the Undiagnosed Diseases Network.

Authors:  Kimberly LeBlanc; Emily G Kelley; Anna Nagy; Jorick Bater; Tala Berro; Molly A McGuinness; Courtney Studwell; Matthew Might
Journal:  Orphanet J Rare Dis       Date:  2021-05-10       Impact factor: 4.123

5.  Unity Is Intelligence: A Collective Intelligence Experiment on ECG Reading to Improve Diagnostic Performance in Cardiology.

Authors:  Luca Ronzio; Andrea Campagner; Federico Cabitza; Gian Franco Gensini
Journal:  J Intell       Date:  2021-04-01

Review 6.  [Digital diagnostic support in rheumatology].

Authors:  J Knitza; M Krusche; J Leipe
Journal:  Z Rheumatol       Date:  2021-10-04       Impact factor: 1.372

7.  Examining Peer-to-Peer and Patient-Provider Interactions on a Social Media Community Facilitating Ask the Doctor Services.

Authors:  Alicia L Nobles; Eric C Leas; Mark Dredze; John W Ayers
Journal:  Proc Int AAAI Conf Weblogs Soc Media       Date:  2020-06

8.  Crowdsourcing Precision Cerebrovascular Health: Imaging and Cloud Seeding A Million Brains Initiative™.

Authors:  David S Liebeskind
Journal:  Front Med (Lausanne)       Date:  2016-11-22

9.  North American Public Opinion Survey on the Acceptability of Crowdsourcing Basic Life Support for Out-of-Hospital Cardiac Arrest With the PulsePoint Mobile Phone App.

Authors:  Katie N Dainty; Haris Vaid; Steven C Brooks
Journal:  JMIR Mhealth Uhealth       Date:  2017-05-17       Impact factor: 4.773

10.  The Impact of an Online Crowdsourcing Diagnostic Tool on Health Care Utilization: A Case Study Using a Novel Approach to Retrospective Claims Analysis.

Authors:  Jessie L Juusola; Thomas R Quisel; Luca Foschini; Joseph A Ladapo
Journal:  J Med Internet Res       Date:  2016-06-01       Impact factor: 5.428

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.