Literature DB >> 32961307

User-centered design of a web-based crowdsourcing-integrated semantic text annotation tool for building a mental health knowledge base.

Xing He1, Hansi Zhang2, Jiang Bian3.   

Abstract

BACKGROUND: One in five U.S. adults lives with some kind of mental health condition and 4.6% of all U.S. adults have a serious mental illness. The Internet has become the first place for these people to seek online mental health information for help. However, online mental health information is not well-organized and often of low quality. There have been efforts in building evidence-based mental health knowledgebases curated with information manually extracted from the high-quality scientific literature. Manual extraction is inefficient. Crowdsourcing can potentially be a low-cost mechanism to collect labeled data from non-expert laypeople. However, there is not an existing annotation tool integrated with popular crowdsourcing platforms to perform the information extraction tasks. In our previous work, we prototyped a Semantic Text Annotation Tool (STAT) to address this gap.
OBJECTIVE: We aimed to refine the STAT prototype (1) to improve its usability and (2) to enhance the crowdsourcing workflow efficiency to facilitate the construction of evidence-based mental health knowledgebase, following a user-centered design (UCD) approach.
METHODS: Following UCD principles, we conducted four design iterations to improve the initial STAT prototype. In the first two iterations, usability testing focus groups were conducted internally with 8 participants recruited from a convenient sample, and the usability was evaluated with a modified System Usability Scale (SUS). In the following two iterations, usability testing was conducted externally using the Amazon Mechanical Turk (MTurk) platform. In each iteration, we summarized the usability testing results through thematic analysis, identified usability issues, and conducted a heuristic evaluation to map identified usability issues to Jakob Nielsen's usability heuristics. We collected suggested improvements in the usability testing sessions and enhanced STAT accordingly in the next UCD iteration. After four UCD iterations, we conducted a case study of the system on MTurk using mental health related scientific literature. We compared the performance of crowdsourcing workers with two expert annotators from two aspects: efficiency and quality.
RESULTS: The SUS score increased from 70.3 ± 12.5 to 81.1 ± 9.8 after the two internal UCD iterations as we improved STAT's functionality based on the suggested improvements. We then evaluated STAT externally through MTurk in the following two iterations. The SUS score decreased to 55.7 ± 20.1 in the third iteration, probably because of the complexity of the tasks. After further simplification of STAT and the annotation tasks with an improved annotation guideline, the SUS score increased to 73.8 ± 13.8 in the fourth iteration of UCD. In the evaluation case study, on average, the workers spent 125.5 ± 69.2 s on the onboarding tutorial and the crowdsourcing workers spent significantly less time on the annotation tasks compared to the two experts. In terms of annotation quality, the workers' annotation results achieved average F1-scores ranged from 0.62 to 0.84 for the different sentences.
CONCLUSIONS: We successfully developed a web-based semantic text annotation tool, STAT, to facilitate the curation of semantic web knowledgebases through four UCD iterations. The lessons learned from the UCD process could serve as a guide to further enhance STAT and the development and design of other crowdsourcing-based semantic text annotation tasks. Our study also showed that a well-organized, informative annotation guideline is as important as the annotation tool itself. Further, we learned that a crowdsourcing task should consist of multiple simple microtasks rather than a complicated task.
Copyright © 2020 Elsevier Inc. All rights reserved.

Entities:  

Keywords:  Crowdsourcing; Mental health; Semantic annotation; Semantic web knowledgebase; User-centered design

Mesh:

Year:  2020        PMID: 32961307      PMCID: PMC7577935          DOI: 10.1016/j.jbi.2020.103571

Source DB:  PubMed          Journal:  J Biomed Inform        ISSN: 1532-0464            Impact factor:   6.317


  12 in total

1.  The interaction of domain knowledge and linguistic structure in natural language processing: interpreting hypernymic propositions in biomedical text.

Authors:  Thomas C Rindflesch; Marcelo Fiszman
Journal:  J Biomed Inform       Date:  2003-12       Impact factor: 6.317

Review 2.  The stigma of mental health problems in the military.

Authors:  Tiffany M Greene-Shortridge; Thomas W Britt; Carl Andrew Castro
Journal:  Mil Med       Date:  2007-02       Impact factor: 1.437

3.  Stigma and help seeking for mental health among college students.

Authors:  Daniel Eisenberg; Marilyn F Downs; Ezra Golberstein; Kara Zivin
Journal:  Med Care Res Rev       Date:  2009-05-19       Impact factor: 3.929

4.  Mental health of young people in the United States: what role can the internet play in reducing stigma and promoting help seeking?

Authors:  Jane M Burns; Lauren A Durkin; Jonathan Nicholas
Journal:  J Adolesc Health       Date:  2009-02-28       Impact factor: 5.012

5.  Wikidata: A large-scale collaborative ontological medical database.

Authors:  Houcemeddine Turki; Thomas Shafee; Mohamed Ali Hadj Taieb; Mohamed Ben Aouicha; Denny Vrandečić; Diptanshu Das; Helmi Hamdi
Journal:  J Biomed Inform       Date:  2019-09-23       Impact factor: 6.317

Review 6.  The quality of mental disorder information websites: a review.

Authors:  Nicola J Reavley; Anthony F Jorm
Journal:  Patient Educ Couns       Date:  2010-11-18

7.  Stigma about depression and its impact on help-seeking intentions.

Authors:  Lisa J Barney; Kathleen M Griffiths; Anthony F Jorm; Helen Christensen
Journal:  Aust N Z J Psychiatry       Date:  2006-01       Impact factor: 5.744

8.  Evaluating the quality of perinatal anxiety information available online.

Authors:  Paige L Kirby; Kristin A Reynolds; John R Walker; Patricia Furer; Teaghan A M Pryor
Journal:  Arch Womens Ment Health       Date:  2018-06-22       Impact factor: 3.633

9.  STAT: A Web-based Semantic Text Annotation Tool to Assist Building Mental Health Knowledge Base.

Authors:  Xing He; Hansi Zhang; Xi Yang; Yi Guo; Jiang Bian
Journal:  IEEE Int Conf Healthc Inform       Date:  2019-11-21

10.  OC-2-KB: integrating crowdsourcing into an obesity and cancer knowledge base curation system.

Authors:  Juan Antonio Lossio-Ventura; William Hogan; François Modave; Yi Guo; Zhe He; Xi Yang; Hansi Zhang; Jiang Bian
Journal:  BMC Med Inform Decis Mak       Date:  2018-07-23       Impact factor: 2.796

View more
  2 in total

1.  Follow-up Interactive Long-Term Expert Ranking (FILTER): a crowdsourcing platform to adjudicate risk for survivorship care.

Authors:  Alex C Cheng; Li Wen; Yanwei Li; Tatsuki Koyama; Lynne D Berry; Tuya Pal; Debra L Friedman; Travis J Osterman
Journal:  JAMIA Open       Date:  2021-11-06

2.  Upper limb rehabilitation system based on virtual reality for breast cancer patients: Development and usability study.

Authors:  Zijun Zhou; Jiaxin Li; He Wang; Ze Luan; Yuan Li; Xin Peng
Journal:  PLoS One       Date:  2021-12-15       Impact factor: 3.240

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.