Literature DB >> 35770149

Using clinical history taking chatbot mobile app for clinical bedside teachings - A prospective case control study.

Michael Co1, Tsz Hon John Yuen2, Ho Hung Cheung1.   

Abstract

Introduction: A novel chatbot mobile app for training of undergraduate medical students' clinical history taking skills was developed in 2021. Students were able to take clinical history from the virtual patient for bedside teaching. A case-control study was conducted to evaluate the effectiveness of learning with chatbot mobile app, versus conventional bedside teachings with real patients.
Methods: 132 final year medical students were randomized into two groups - Conventional bedside teaching with clinical history taken from a real patient, and Bedside teaching with clinical history taken from the Chatbot. Independent blinded assessment of students' history taking skills was conducted. Students' performance were assessed by standardized marking scheme.
Results: Median age was 23 years old (Range 21-30 years old). There were 62 female and 70 male students.64 students were randomized into conventional group while 68 students were randomized into the chatbot group. Baseline demographic data were comparable between the two groups.Blinded assessment of students' performance in clinical history taking were comparable between the conventional group and chatbot group (p > 0.05).
Conclusion: With the promising results we have demonstrated in this study, we believe training of history taking skills by chatbot will be a feasible alternative to conventional bedside teaching.
© 2022 The Authors. Published by Elsevier Ltd.

Entities:  

Keywords:  Clinical education; Computers; New technology; Simulation

Year:  2022        PMID: 35770149      PMCID: PMC9234591          DOI: 10.1016/j.heliyon.2022.e09751

Source DB:  PubMed          Journal:  Heliyon        ISSN: 2405-8440


Introduction

Undergraduate medical education has been severely affected by the COVID-19 outbreak. Many clinical teaching activities were replaced by virtual or online formats, which were subsequently published on peer-reviewed journals [1, 2]. While lectures can be easily replaced by pre-recorded videos or delivered live by telecommunication software (e.g. Zoom), clinical bedside teachings, including training of history taking skills from patient cannot be easily replaced by online teaching. Conventional clinical history taking skills were acquired through medical student-patient interaction. Medical students were allowed to visit hospital wards and talk to the patients. Usually, students were required to report and summarize the clinical history to the tutor afterwards where comments will be given (i.e. Conventional bedside clinical teaching). During COVID-19 pandemic in Hong Kong, medical students were not able to visit clinical areas due to infection risk. As such, medical students were not able to see patients in wards. Chatbot app has been widely used in many commercial companies like banks, hotels and airline industries, its application in training medical students’ history taking skills has not been described in the medical education community. As such, we have developed a novel clinical history taking Chatbot app (Bennie and the Chats) for bedside clinical teaching, which has been rolled out to final year undergraduate medical students in our institution since March 2021. Medical students were asked to download the Chatbot app to their own mobile device (e.g. mobile phone or tablet). They can then ask questions directly to the Chatbot and collect clinical history by typing in questions. At instance, voice recognition function is not available in this Chatbot mobile app, students will need to type in questions manually. Nevertheless, medical students were able to interact with the virtual patient Bennie (The chatbot) and take clinical history by asking questions. Chatbot Bennie will give relevant answers according to the pre-defined keyword database designed by the tutors (MC & BC). Students were able to discuss the case scenario and report the clinical history to the tutor (MC & BC) as in conventional bedside clinical teaching in pre-COVID-19 era. This novel teaching method was introduced in March 2021 and here we present our initial experience with using chatbot mobile app for training of clinical history taking skills in final year undergraduate medical students.

Materials and methods

Development of the “chatbot app” (also named Bennie and the chats)

The Chatbot app (Bennie and the chats) was co-developed by authors M Co and TH Yuen in 2020, using a number of software from Google. At the front end, the student input is handled by “Actions on Google” in a conversational user interface. The input message from the student is processed by Dialogflow to understand the meaning of the questions. Dialogflow is a natural language processing platform which uses the machine learning technology to understand how people talk and it extracts the intent in a conversation. Afterwards, Dialogflow will find a suitable answer from a database. In “Actions on Google”, user interface is designed using simple flow diagram. User input on the interface can be detected and the transitions between different stages can be easily designed over the Actions Console. The app is equipped with artificial intelligence and natural language processing in which consistent answers will be given from the chatbot regardless of the words/phrases used by the student. For example, when a student asks “What is your main complaint?” the app will answer the pre-defined phrase, for example “I have a right breast lump” (Figure 1). Even when another student enters another question such as “What is your problem?” The chatbot will still answer “I have right breast lump”. Dialogflow enables developer to define synonyms for all keywords used in this app.
Figure 1

Two separate screen captures of conversation between medical student and Chatbot Bennie, illustrating the student-chatbot interaction (Right) and chatbot's response to synonym (Left).

Two separate screen captures of conversation between medical student and Chatbot Bennie, illustrating the student-chatbot interaction (Right) and chatbot's response to synonym (Left). At the back end, we implemented the fulfilment that contains the logic to construct a dynamic conversational response. The Actions Console can connect the apps to the cloud function through the Webhook interface. This Chatbot app is pre-installed with case scenarios from different surgical subspecialties, including Breast Surgery, Colorectal surgery, Endocrine Surgery, Esophageal and upper gastrointestinal surgery, Head and Neck Surgery, Hepatobiliary Surgery, Urology, and Vascular Surgery.

Introduction of chatbot app to medical curriculum

The undergraduate medical curriculum in Hong Kong is composed of three-year pre-clinical teaching followed by another three-year clinical teaching. From 4th year on, medical students are required to acquire clinical history taking and physical examination skills. Conventional teaching of clinical history taking involves face-to-face patient clerking. Medical Students are required to take clinical history from the patient followed by presentation of medical history to the assigned clinical tutors in the ward. Chatbot app was rolled out to the final year medical students in Hong Kong in 2021. Students were advised to take clinical history from the virtual patient in the Chatbot, followed by interactive tutorial with the assigned tutor.

Research question and methodology

This study aims to evaluate the feasibility and efficacy of bedside clinical teaching using chatbot app, based on students’ performance on clinical history presentation. Approval was sought from Center for Education and Training, Department of Surgery, University of Hong Kong for the curriculum adjustments. 132 final year medical students were recruited. Students were randomized into two groups – (A) Conventional bedside teaching with clinical history taken from a real patient, and (B) Bedside teaching with clinical history taken from the Chatbot (Virtual patient Bennie). Independent blinded assessment of students’ history taking skills was conducted. All students were given 45 min to gather history from either genuine patient or from the chatbot app (virtual patient), followed by 45 min discussion on the clinical scenario with assigned tutor. Clinical history taken by students were assessed based on the comprehensiveness in 10 aspects of history taking – Chief complaint, duration of symptom, history of present illness, associated symptoms, past medical history, past surgical history, family history, risk factors, investigations performed, and social history. Number of students who were unable to obtain information in each individual aspect were recorded. Students who were randomized to the Chatbot group were invited to rate the chatbot system based on user friendliness, accuracy of keyword identification, student-chatbot interaction, efficiency of learning and overall learning experience in a Likert scale of 1–10.

Results

A total of 132 medical students were recruited into the study, median age was 23 years old (Range 21–30 years old). There were 62 female and 70 male students. All students were in their final year of undergraduate medical curriculum. 64 students were randomized into conventional group while 68 students were randomized into the chatbot group. Median age was 22 (21–30) in the conventional group while that in the chatbot group was 23 (21–26) (p = 0.31). 5 (7.8%) students from the conventional group and 6 (8.8%) students from the chatbot group were post-graduate students (i.e. who have already obtained another bachelor or above degree) (p = 0.83). 4 (6.3%) students from the conventional group and 4 (5.9%) students from the chatbot group had history of distinction (merit) award in previous examinations (p = 0.93). 5 (7.8%) students from the conventional group and 5 (7.4%) students from the chatbot group had history of failure in previous examinations (p = 0.92). Baseline demographic data were summarized in Table 1.
Table 1

Baseline demographic data of each group of students.

Conventional bedsideN = 64ChatbotN = 68P-value
Median age (Range)22 (21–30)23 (21–26)0.31
Male gender32 (50%)38 (55.9%)0.50
Post-graduate5 (7.8%)6 (8.8%)0.83
Distinction (Merit) in previous exams4 (6.3%)4 (5.9%)0.93
Failure in previous exams5 (7.8%)5 (7.4%)0.92
Baseline demographic data of each group of students. Students’ performance in clinical history taking were comparable between the conventional group and chatbot group (Table 2). All students from either group were able to obtain accurate clinical history on Chief complaint, history of present illness, duration of symptoms, history of present illness, past medical history and family history. 8 (12.5%) students from the conventional group and 12 (17.6%) students from the chatbot group were unable to obtain accurate surgical history (p = 0.41). 2 (3.1%) students from the conventional group and 2 (2.9%) students from the chatbot group were unable to obtain accurate risk factors (p = 0.95). 6 (9.4%) students from the conventional group and 6 (8.8%) students from the chatbot group were unable to obtain history on investigations (p = 0.91), 2 (3.2%) students from the conventional group and 3 (4.7%) students from the chatbot group were unable to obtain accurate surgical history (p = 0.70).
Table 2

Number of students who were unable to take the correct history in each group.

Conventional bedside (N = 64)Chatbot (N = 68)p-value
Chief complaint0 (0%)0 (0%)1.00
Duration of symptoms0 (0%)0 (0%)1.00
History of present illness0 (0%)0 (0%)1.00
Associated symptoms0 (0%)0 (0%)1.00
Past medical history0 (0%)0 (0%)1.00
Past surgical history8 (12.5%)12 (17.6%)0.41
Family history0 (0%)0 (0%)1.00
Risk factors2 (3.1%)2 (2.9%)0.95
Investigations6 (9.4%)6 (8.8%)0.91
Social history2 (3.2%)3 (4.7%)0.70
Overall14 (21.9%)22 (32.4%)0.18
Number of students who were unable to take the correct history in each group. Students’ feedback on clinical history taking from the chatbot system were generally positive. Median Likert scores of user friendliness, keyword identification, student-chatbot interaction, efficiency of learning and overall experience were 8 (Range 6–10), 7 (Range 5–9), 7 (Range 5–9), 8 (Range 6–9) and 8 (Range 6–9), respectively (Table 3).
Table 3

Students’ evaluation of the Bernie Chatbot System (Likert scale of 1–10).

Areas of evaluationMedian students rating out of 10 (Range)
User friendliness8 (6–10)
Keyword identification7 (5–9)
Interaction7 (6–9)
Efficiency of learning8 (6–9)
Overall learning experience8 (6–9)
Students’ evaluation of the Bernie Chatbot System (Likert scale of 1–10).

Discussion

Distant learning has been widely adopted in teaching of medical students during COVID-19 pandemic. Medical students were not allowed to enter hospital premises during the episodes of the outbreak. Lectures were replaced by pre-recorded videos, clinical skills were replaced by live demonstration [1, 2]. However, learning of history taking skills, which involves student-patient interaction, cannot be easily replaced by online lectures. This has resulted significant stress to undergraduate medical students [3]. Prior to the pandemic. Medical students learnt and practiced clinical history taking by talking to the patients in the wards or clinics. COVID-19 pandemic has given the opportunity to explore alternative ways for clinical history clerking for medical students when face-to-face clinical teaching was not allowed [4, 5]. For example, simulated patients have been used to train communication skills in health professional education [6, 7]. Computer chatbot has been widely used in commercial companies, it has also been described as an online tutor for medical students [8, 9]. However, the use of a chatbot system in facilitating clinical history clerking among medical students has never been described. This new training method for clinical history taking will allow efficient training of medical students’ history taking skills without geographical and time constrain. Contents of the mobile app will also be updated regularly. The aim of this mobile app is to allow training of history taking skills among medical students by conversations with the virtual patient. Distant learning will remain as an important teaching modality for medical students even when the pandemic is over [10, 11, 12]. Medical students will be able to take clinical history from the virtual patient Bennie (Chatbot mobile app) without time or geographical restriction. In addition, students will be able to interact with chatbots with pre-recorded rare clinical scenarios, which is not commonly seen in daily clinical practice. This study was conducted on final year medical students, most of them have already acquired necessary clinical history taking skills from year 4 of the medical curriculum. However, the aim of the study is to evaluate the feasibility of using chatbot app for clinical bedside teaching. All students were able to gather necessary clinical history from the chatbot app, meaning that clinical teaching with chatbot app is a feasible alternative to conventional bedside teaching on real patients. A pilot questionnaire on students' satisfaction on the chatbot app was adopted in this study. We do recognize this as a potential limitation of this study. System usability scale (SUS) could be a validated alternative to the pilot questionnaire that was used in this study [13]. However, SUS is a generic questionnaire and does not evaluate on specific areas of concerns such as students’ satisfaction on keyword identification and chatbot-user interaction. Nevertheless, feedbacks from medical students have been positive, although some students reflected on the limited keyword identification and student-chatbot by the Chatbot system. These can be overcome by updating the keyword database of the Chatbot system.

Conclusion

With the promising results we have demonstrated in this study, we believe training of history taking skills by chatbot will be a feasible alternative to conventional bedside teaching.

Declarations

Author contribution statement

Michael Co: Conceived and designed the experiments; Performed the experiments; Analyzed and interpreted the data; Contributed reagents, materials, analysis tools or data; Wrote the paper. Tsz Hon John Yuen: Conceived and designed the experiments; Performed the experiments; Contributed reagents, materials, analysis tools or data; Wrote the paper. Ho Hung Cheung: Conceived and designed the experiments; Performed the experiments; Analyzed and interpreted the data.

Funding statement

This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

Data availability statement

Data will be made available on request.

Declaration of interests statement

The authors declare no conflict of interest. Dr. Michael Co serves as an unpaid Guest Editor of the Journal. He is not involved in the peer-review nor editorial process of this article.

Additional information

No additional information is available for this paper.
  11 in total

1.  Impact of the COVID-19 pandemic on medical education: Medical students' knowledge, attitudes, and practices regarding electronic learning.

Authors:  Ahmed Alsoufi; Ali Alsuyihili; Ahmed Msherghi; Ahmed Elhadi; Hana Atiyah; Aimen Ashini; Arwa Ashwieb; Mohamed Ghula; Hayat Ben Hasan; Salsabil Abudabuos; Hind Alameen; Taqwa Abokhdhir; Mohamed Anaiba; Taha Nagib; Anshirah Shuwayyah; Rema Benothman; Ghalea Arrefae; Abdulwajid Alkhwayildi; Abdulmueti Alhadi; Ahmed Zaid; Muhammed Elhadi
Journal:  PLoS One       Date:  2020-11-25       Impact factor: 3.240

2.  Medical education challenges and innovations during COVID-19 pandemic.

Authors:  Michail Papapanou; Eleni Routsi; Konstantinos Tsamakis; Lampros Fotis; Georgios Marinos; Irene Lidoriki; Marianna Karamanou; Theodore G Papaioannou; Dimitrios Tsiptsios; Nikolaos Smyrnis; Emmanouil Rizos; Dimitrios Schizas
Journal:  Postgrad Med J       Date:  2021-03-29       Impact factor: 2.401

Review 3.  Distance education for anatomy and surgical training - A systematic review.

Authors:  Michael Co; Kwong Yee Chloe Cheung; Wan Suen Cheung; Ho Man Fok; Ka Hey Fong; Oi Yan Kwok; Tsz Wai Kelvin Leung; Hei Chun Justin Ma; Pui Ting Isabelle Ngai; Man Kit Tsang; Cheuk Yin Matthew Wong; Kent-Man Chu
Journal:  Surgeon       Date:  2021-09-03       Impact factor: 2.632

4.  The Role for Virtual Patients in the Future of Medical Education.

Authors:  Norman B Berman; Steven J Durning; Martin R Fischer; Soren Huwendiek; Marc M Triola
Journal:  Acad Med       Date:  2016-09       Impact factor: 6.893

5.  Distant surgical teaching during COVID-19 - A pilot study on final year medical students.

Authors:  Michael Co; Kent-Man Chu
Journal:  Surg Pract       Date:  2020-07-10

6.  The impact of COVID-19 on medical education: Experiences from one medical university in Taiwan.

Authors:  Chih-Hao Liu; Hugo You-Hsien Lin
Journal:  J Formos Med Assoc       Date:  2021-03-08       Impact factor: 3.282

7.  How medical education survives and evolves during COVID-19: Our experience and future direction.

Authors:  Ju Whi Kim; Sun Jung Myung; Hyun Bae Yoon; Sang Hui Moon; Hyunjin Ryu; Jae-Joon Yim
Journal:  PLoS One       Date:  2020-12-18       Impact factor: 3.240

8.  Online teaching of basic surgical skills to medical students during the COVID-19 pandemic: a case-control study.

Authors:  Michael Co; Patrick Ho-Yu Chung; Kent-Man Chu
Journal:  Surg Today       Date:  2021-01-25       Impact factor: 2.549

9.  Qualitative exploration of digital chatbot use in medical education: A pilot study.

Authors:  Anjuli Kaur; Sid Singh; Joht S Chandan; Tim Robbins; Vinod Patel
Journal:  Digit Health       Date:  2021-08-31

10.  The utilization of simulated patients for teaching and learning in the pharmacy curriculum: exploring pharmacy students' and recent alumni's perceptions using mixed-methods approach.

Authors:  Hager ElGeed; Maguy Saffouh El Hajj; Raja Ali; Ahmed Awaisu
Journal:  BMC Med Educ       Date:  2021-11-06       Impact factor: 2.463

View more
  1 in total

1.  Using a Virtual Patient via an Artificial Intelligence Chatbot to Develop Dental Students' Diagnostic Skills.

Authors:  Ana Suárez; Alberto Adanero; Víctor Díaz-Flores García; Yolanda Freire; Juan Algar
Journal:  Int J Environ Res Public Health       Date:  2022-07-18       Impact factor: 4.614

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.