Literature DB >> 31661083

Trust Me, I'm a Chatbot: How Artificial Intelligence in Health Care Fails the Turing Test.

John Powell1.   

Abstract

Over the next decade, one issue which will dominate sociotechnical studies in health informatics is the extent to which the promise of artificial intelligence in health care will be realized, along with the social and ethical issues which accompany it. A useful thought experiment is the application of the Turing test to user-facing artificial intelligence systems in health care (such as chatbots or conversational agents). In this paper I argue that many medical decisions require value judgements and the doctor-patient relationship requires empathy and understanding to arrive at a shared decision, often handling large areas of uncertainty and balancing competing risks. Arguably, medicine requires wisdom more than intelligence, artificial or otherwise. Artificial intelligence therefore needs to supplement rather than replace medical professionals, and identifying the complementary positioning of artificial intelligence in medical consultation is a key challenge for the future. In health care, artificial intelligence needs to pass the implementation game, not the imitation game. ©John Powell. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 28.10.2019.

Entities:  

Keywords:  artificial intelligence; chatbots; conversational agents; digital health; ehealth; machine learning; medical informatics

Mesh:

Year:  2019        PMID: 31661083      PMCID: PMC6914236          DOI: 10.2196/16222

Source DB:  PubMed          Journal:  J Med Internet Res        ISSN: 1438-8871            Impact factor:   5.428


Over the last two decades, the concerns of digital health researchers interested in the social impact of the internet have evolved as the technology has matured and new tools have emerged. From a sociotechnical perspective, there were initial preoccupations with the impact of a new, uncontrolled form of mass communication, alongside concerns with the quality of unregulated online information and threats to professions, with medical professionals in particular fearing a loss of authority [1-3]. As Web2.0 developments took hold and the public became producers as well as consumers of health information, researchers began to identify the benefits of online peer-to-peer communication and the sharing of information in virtual communities, social media, and increasingly on health ratings sites [4-7]. With the mass uptake in smartphones, the subsequent rapid developments in mobile health, and the explosion in health apps, we are now exploring the value of low-cost, patient-centered interventions delivered directly to consumers [8,9]. In addition, we are also gaining a better understanding of the limitations and key issues in their implementation, such as nonadoption and abandonment [10]. As the number one journal in this field, the Journal of Medical Internet Research continues to reflect and illuminate all these debates. For those of us studying the social science of digital technology in health and health care, one area of research is likely to dominate the next decade: the extent to which the promise of artificial intelligence (AI) in health care will be realized, and the social and ethical issues which accompany it [11-13]. Broadly speaking, we can identify two current strands in the use of AI in health care. Firstly, there are data-facing applications which use techniques such as machine learning and artificial neural networks to derive new knowledge from large datasets, such as improving diagnostic accuracy from scans and other images [14]. Secondly, there are user-facing applications and intelligent agents which interact with people in real-time, using inferences to provide advice or instruction based on probabilities which the tool can derive and improve over time, such as a chatbot substituting or complementing a health care consultation with a patient [15]. In this article I focus on the latter to consider the approaches of these chatbots, or “robot doctors,” to medical consultation, and specifically the extent to which these technologies will ever pass the celebrated Turing test. Alan Turing, the British mathematician and theoretical computer scientist, is widely regarded as the founding father of AI. He proposed that for a machine to be considered intelligent it should provide responses to a blinded interrogation that are indistinguishable from those given by a human comparator [16]. In other words, the interrogator should not be able to tell whether the machine or the human was responding. If we extrapolate this thought experiment to current health care, we can pose the question of whether AI-based medical consultations (conversational agents and medical chatbots) will ever be considered intelligent by Turing’s standard. Of course, context is important, and if a patient is asking a simple factual question that requires a binary response, for example, then even current AI systems can mimic a human interlocutor with high accuracy. However, we know that medical consultations are complex [17], that many medical decisions require value judgements, and that the doctor-patient relationship requires empathy and understanding to arrive at a shared decision [18]. The practice of medicine is as much an art as a science, and patients may choose a path which is not necessarily the one that logic would determine. Even the pioneers of evidence-based medicine defined their normative approach as: the conscientious and judicious use of current best evidence from clinical care research in the management of individual patients [19]. Conscience and the ability to weigh competing personal values are not strengths of AI. A key skill for medical professionals is the ability to deal with uncertainty alongside considering patients’ preferences. What doctors often need is wisdom rather than intelligence, and we are a long way away from a science of artificial wisdom. It is doubtful whether AI will ever pass the Turing test for complex medical consultations, but this is to misunderstand the place of AI in future medical care. AI should complement rather than replace medical professionals. As various studies into the future of work have shown, automation in the workplace will not eliminate all human tasks [20]. Chatbot approaches have many potential benefits, including the potential to allow clinicians to have more time for delivering empathic and personalized care [15]. Perhaps, as a senior clinical informatics leader in the UK has suggested, “AI will allow doctors to be more human” [13]. However, as has been well established for many innovations in health care, especially digital ones, the key challenges for health systems seeking to harness the benefits of the technology are not just related to its effectiveness but also to the wider issues of its integration and implementation [10,12,21]. We need to understand how to integrate the tools and practices of AI within the work and culture of professionals and organizations, to investigate factors related to adoption, nonadoption, and abandonment [10,12], and investigate the work required to sustain innovation [22]. Factors which will influence the implementation of AI tools include those related to people, such as professional and public attitudes, trust, existing work practices, training needs, and the risks of deskilling and disempowerment; those related to the health system, such as leadership and management, the positioning of clinical responsibility and accountability, and the possibility of harm, alongside issues of regulation and service provision (including scalability and the possibility of providing two-tier services with or without AI); those related to the data, such as issues of data security, privacy, consent and ownership; and those related to the tool itself, such as transparency of the algorithm, issues of reliability and validity, and algorithmic bias [12,21,23]. To take an example, in an early study of an algorithm-based triage tool in primary care, we showed that physicians lacked trust in the ability of the machine to take clinical risks and worried about issues of governance and accountability, such that the sensitivity of the tool, in terms of the urgency of triage, was consistently set at a threshold which would increase urgent clinical workload rather than reduce it [24]. Identifying the complementary positioning of AI tools in health care in general, and in particular for their use in the medical consultation, is a key challenge for the future. We need to understand how to integrate the precision and power of AI tools and practices with the wisdom and empathy of the doctor-patient relationship. In health care, it is more important that artificial intelligence passes the implementation game rather than the imitation game.
  19 in total

Review 1.  Complex consultations and the 'edge of chaos'.

Authors:  Andrew D Innes; Peter D Campion; Frances E Griffiths
Journal:  Br J Gen Pract       Date:  2005-01       Impact factor: 5.386

2.  Evidence based medicine: what it is and what it isn't.

Authors:  D L Sackett; W M Rosenberg; J A Gray; R B Haynes; W S Richardson
Journal:  BMJ       Date:  1996-01-13

3.  A cross sectional survey of the UK public to understand use of online ratings and reviews of health services.

Authors:  Michelle H van Velthoven; Helen Atherton; John Powell
Journal:  Patient Educ Couns       Date:  2018-04-09

4.  The importance of being expert: the quest for cancer information on the Internet.

Authors:  Sue Ziebland
Journal:  Soc Sci Med       Date:  2004-11       Impact factor: 4.634

5.  Ten key considerations for the successful implementation and adoption of large-scale health information technology.

Authors:  Kathrin M Cresswell; David W Bates; Aziz Sheikh
Journal:  J Am Med Inform Assoc       Date:  2013-04-18       Impact factor: 4.497

Review 6.  Assessing the Efficacy of Mobile Health Apps Using the Basic Principles of Cognitive Behavioral Therapy: Systematic Review.

Authors:  Amy Leigh Rathbone; Laura Clarry; Julie Prescott
Journal:  J Med Internet Res       Date:  2017-11-28       Impact factor: 5.428

Review 7.  Artificial Intelligence Versus Clinicians in Disease Diagnosis: Systematic Review.

Authors:  Jiayi Shen; Casper J P Zhang; Bangsheng Jiang; Jiebin Chen; Jian Song; Zherui Liu; Zonglin He; Sum Yi Wong; Po-Han Fang; Wai-Kit Ming
Journal:  JMIR Med Inform       Date:  2019-08-16

8.  Artificial Intelligence and the Implementation Challenge.

Authors:  James Shaw; Frank Rudzicz; Trevor Jamieson; Avi Goldfarb
Journal:  J Med Internet Res       Date:  2019-07-10       Impact factor: 5.428

9.  Using computer decision support systems in NHS emergency and urgent care: ethnographic study using normalisation process theory.

Authors:  Catherine Pope; Susan Halford; Joanne Turnbull; Jane Prichard; Melania Calestani; Carl May
Journal:  BMC Health Serv Res       Date:  2013-03-23       Impact factor: 2.655

10.  Cross-sectional survey of users of Internet depression communities.

Authors:  John Powell; Noel McCarthy; Gunther Eysenbach
Journal:  BMC Psychiatry       Date:  2003-12-10       Impact factor: 3.630

View more
  10 in total

1.  Self-Diagnosis through AI-enabled Chatbot-based Symptom Checkers: User Experiences and Design Considerations.

Authors:  Yue You; Xinning Gui
Journal:  AMIA Annu Symp Proc       Date:  2021-01-25

2.  2021 ISHNE/HRS/EHRA/APHRS Expert Collaborative Statement on mHealth in Arrhythmia Management: Digital Medical Tools for Heart Rhythm Professionals: From the International Society for Holter and Noninvasive Electrocardiology/Heart Rhythm Society/European Heart Rhythm Association/Asia-Pacific Heart Rhythm Society.

Authors:  Niraj Varma; Iwona Cygankiewicz; Mintu P Turakhia; Hein Heidbuchel; Yu-Feng Hu; Lin Yee Chen; Jean-Philippe Couderc; Edmond M Cronin; Jerry D Estep; Lars Grieten; Deirdre A Lane; Reena Mehra; Alex Page; Rod Passman; Jonathan P Piccini; Ewa Piotrowicz; Ryszard Piotrowicz; Pyotr G Platonov; Antonio Luiz Ribeiro; Robert E Rich; Andrea M Russo; David Slotwiner; Jonathan S Steinberg; Emma Svennberg
Journal:  Circ Arrhythm Electrophysiol       Date:  2021-02-12

3.  2021 ISHNE/HRS/EHRA/APHRS Collaborative Statement on mHealth in Arrhythmia Management: Digital Medical Tools for Heart Rhythm Professionals: From the International Society for Holter and Noninvasive Electrocardiology/Heart Rhythm Society/European Heart Rhythm Association/Asia Pacific Heart Rhythm Society.

Authors:  Niraj Varma; Iwona Cygankiewicz; Mintu P Turakhia; Hein Heidbuchel; Yufeng Hu; Lin Yee Chen; Jean-Philippe Couderc; Edmond M Cronin; Jerry D Estep; Lars Grieten; Deirdre A Lane; Reena Mehra; Alex Page; Rod Passman; Jonathan P Piccini; Ewa Piotrowicz; Ryszard Piotrowicz; Pyotr G Platonov; Antonio Luiz Ribeiro; Robert E Rich; Andrea M Russo; David Slotwiner; Jonathan S Steinberg; Emma Svennberg
Journal:  Cardiovasc Digit Health J       Date:  2021-01-29

4.  Caregiver Expectations of Interfacing With Voice Assistants to Support Complex Home Care: Mixed Methods Study.

Authors:  Ryan Tennant; Sana Allana; Kate Mercer; Catherine M Burns
Journal:  JMIR Hum Factors       Date:  2022-06-30

5.  Quantitative User Data From a Chatbot Developed for Women With Gestational Diabetes Mellitus: Observational Study.

Authors:  Mari Haaland Sagstad; Nils-Halvdan Morken; Agnethe Lund; Linn Jannike Dingsør; Anne Britt Vika Nilsen; Linn Marie Sorbye
Journal:  JMIR Form Res       Date:  2022-04-18

Review 6.  Special Section on Ethics in Health Informatics.

Authors:  Carolyn Petersen; Vignesh Subbian
Journal:  Yearb Med Inform       Date:  2020-08-21

Review 7.  Artificial intelligence for good health: a scoping review of the ethics literature.

Authors:  Kathleen Murphy; Erica Di Ruggiero; Ross Upshur; Donald J Willison; Neha Malhotra; Jia Ce Cai; Nakul Malhotra; Vincci Lui; Jennifer Gibson
Journal:  BMC Med Ethics       Date:  2021-02-15       Impact factor: 2.652

8.  Chatbot breakthrough in the 2020s? An ethical reflection on the trend of automated consultations in health care.

Authors:  Jaana Parviainen; Juho Rantala
Journal:  Med Health Care Philos       Date:  2021-09-04

9.  Evaluation of Deep Learning-Based Automated Detection of Primary Spine Tumors on MRI Using the Turing Test.

Authors:  Hanqiang Ouyang; Fanyu Meng; Jianfang Liu; Xinhang Song; Yuan Li; Yuan Yuan; Chunjie Wang; Ning Lang; Shuai Tian; Meiyi Yao; Xiaoguang Liu; Huishu Yuan; Shuqiang Jiang; Liang Jiang
Journal:  Front Oncol       Date:  2022-03-11       Impact factor: 6.244

10.  Celebrating 20 Years of Open Access and Innovation at JMIR Publications.

Authors:  Gunther Eysenbach
Journal:  J Med Internet Res       Date:  2019-12-23       Impact factor: 5.428

  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.