Literature DB >> 35983102

Attitudes and perception of artificial intelligence in healthcare: A cross-sectional survey among patients.

Sebastian J Fritsch1,2,3, Andrea Blankenheim1, Alina Wahl1, Petra Hetfeld1,2, Oliver Maassen1,2, Saskia Deffge1,2, Julian Kunze2,4, Rolf Rossaint4, Morris Riedel2,3,5, Gernot Marx1,2, Johannes Bickenbach1,2.   

Abstract

Objective: The attitudes about the usage of artificial intelligence in healthcare are controversial. Unlike the perception of healthcare professionals, the attitudes of patients and their companions have been of less interest so far. In this study, we aimed to investigate the perception of artificial intelligence in healthcare among this highly relevant group along with the influence of digital affinity and sociodemographic factors.
Methods: We conducted a cross-sectional study using a paper-based questionnaire with patients and their companions at a German tertiary referral hospital from December 2019 to February 2020. The questionnaire consisted of three sections examining (a) the respondents' technical affinity, (b) their perception of different aspects of artificial intelligence in healthcare and (c) sociodemographic characteristics.
Results: From a total of 452 participants, more than 90% already read or heard about artificial intelligence, but only 24% reported good or expert knowledge. Asked on their general perception, 53.18% of the respondents rated the use of artificial intelligence in medicine as positive or very positive, but only 4.77% negative or very negative. The respondents denied concerns about artificial intelligence, but strongly agreed that artificial intelligence must be controlled by a physician. Older patients, women, persons with lower education and technical affinity were more cautious on the healthcare-related artificial intelligence usage. Conclusions: German patients and their companions are open towards the usage of artificial intelligence in healthcare. Although showing only a mediocre knowledge about artificial intelligence, a majority rated artificial intelligence in healthcare as positive. Particularly, patients insist that a physician supervises the artificial intelligence and keeps ultimate responsibility for diagnosis and therapy.
© The Author(s) 2022.

Entities:  

Keywords:  Artificial intelligence; algorithms; attitude; clinical decision support systems; digital divide; digital technology; patients; perception; surveys and questionnaires

Year:  2022        PMID: 35983102      PMCID: PMC9380417          DOI: 10.1177/20552076221116772

Source DB:  PubMed          Journal:  Digit Health        ISSN: 2055-2076


Introduction

Artificial intelligence (AI) is a topic that has become increasingly relevant in the social debate in recent years. Politicians, economists, scientists as well as lay people are talking controversially about this special subject. However, the level of public knowledge about AI is frequently low and its perception is not exclusively positive. While a majority has optimistic opinions about AI’s capability to improve human life, there is also controversial discussion about ethical concerns, loss of control and undesired consequences by uncritical usage of AI. These ambiguous feelings can be observed in the very sensitive areas of the healthcare sector like under a magnifying lens.

Background: AI

Although definitions for AI are as diverse as the definition of intelligence in general, an early AI definition, which was given by the Dartmouth Research Project in 1955, is still valid today: ‘Making a machine behave in ways that would be called intelligent if a human were so behaving’. More in detail, it can be defined as a system's ability to interpret external data correctly, to learn from such data and to use those learnings to achieve specific goals and tasks through flexible adaptation. This quality allows AI to find patterns and subtle and complex associations in large, high-dimensional data that often escape traditional analysis techniques as well. Thus, typical use cases for AI in medicine are image analysis tasks, not only in radiology, but also in ophthalmology, dermatology and pathology. However, also, in clinical risk prediction, diagnostics and therapeutics, the usage of AI allows the consideration of more and more data. For instance, using AI allows the integration of genomic, proteomic and radiomic data to predict cancer outcomes, which is able to increase prediction accuracy relevantly. Finally, the ability to process a higher amount of data in a shorter time period makes it possible to supplement the limited processing power of the human brain and thus reducing the workload of healthcare professionals. Despite the high potential of AI in medicine, however, there are concerns among the stakeholders about the safety of AI but also data security. Patients fear that they might not have the choice to refuse an AI usage for their personal treatment, rising costs and problems with insurance coverage. In a more technical view, missing, erroneous or insufficiently annotated training data may impair the performance of AI models and prevent them to generalize well beyond their original population. Especially biased data result in a biased output and can lead to a discrimination of the underrepresented subgroup. However, also legal issues, uncertainties about privacy and liability and a poor explainability due to the ‘black box’ nature of many AI models result in mistrust and hamper clinical implementation.[9-11]

AI perception of healthcare professionals and patients

During the last few years, numerous researchers focused on healthcare professionals when examining the perception of AI in medicine. For instance, in recent surveys conducted in Germany, France and the United Kingdom, healthcare professionals indicated that their attitude towards AI is generally positive and they expect it to improve their daily work. Nevertheless, they are also aware of the issues described above.[12-14] However, it must be stated that for a reasonable use of AI in healthcare, the acceptance of patients and their families is necessary as well. Modern healthcare aims for participation and cooperation of patients, often described under the term ‘patient empowerment’. Insufficient acceptance of therapeutic measures impairs patient’s compliance and worsens an otherwise possible successful outcome, so concerns about AI could impair the dissemination and use of these tools relevantly.

Current status of the literature

AI in healthcare represents one of the fastest growing research subjects in current medical research. Within the last few years, the number of publications showed a nearly exponential increase. A comprehensive description of this highly dynamic and rapidly growing literature field is challenging and exceeds the bounds of this paper. Reference is therefore made to corresponding publications.[9,16-18] A rough summary of the literature shows that vast majority of the work is in the experimental stage and an implementation in clinical routine has only taken place in exceptions so far. However, many publications state that their AI application would achieve a non-inferior or even better performance compared to human physicians, although poor reporting and high risk of bias is observed frequently in the respective publications. Regardless of the high number of publications, patients’ opinions and general perceptions of AI in healthcare were much less in focus. The results vary relevantly depending on the surveyed group and the AI application in focus. Aggarwal and colleagues examined perceptions of 408 participants from London regarding AI along with questions about data collection and data sharing for the purpose of medical AI research. They found that while patients generally have little prior knowledge about AI, the majority perceives its usage as positive, trusts its use and believes that the benefits outweigh the risks. A common finding in many studies is that patients want physician supervision of AI and prefer a physician in a direct comparison. Thus, a majority of 229 German patients preferred physicians over AI for all clinical tasks. As the only exception, for treatment planning based on current scientific evidence, an AI with physician supervision was strongly preferred. However, nearly all patients favoured a human physician if AI and physician came to differing estimations. Similarly, in a study from the UK involving 107 neurosurgery patients, proportions of two-thirds up to three-fourths accepted AI for clinical tasks if a neurosurgeon was continuously in control. A particularly striking example of the high respect and trust in the competence of physicians compared to AI tools is provided by the study of York and colleagues using the example of radiographic fracture identification. On a 10-point scale representing the confidence, the study population of 216 respondents awarded their human radiologists a near maximum score of 9.2 points, while the AI tool received only 7.0 points. A multitude of additional studies showed that that people are reluctant to trust AI technology and prefer human physicians, even if the performance of the AI system is equal or better.[25-28] In contrast, in a study from the US including 804 parents of paediatric patients, openness to AI-driven tools was given in majority if accuracy was proven and several other quality measures were fulfilled. In contrast, from 1183 mostly female patients with chronic diseases, who lived in France, only 20% expected AI to be beneficial and 35% declined AI usage in their personal care completely. In a dermatological setting, 48 patients from the US see AI usage mainly as a tool for second opinions for their treating physicians. These patients stated that AI exhibits strengths (like continuous high accuracy) as well as weaknesses (rare but serious misdiagnoses) at the same time. As relevant risk factors, an impairment of the relation between patients and physicians, untrustworthy AI and missing regulation was indicated in a US online survey. This perceived risk even persisted in an experimental setting among 634 respondents from the US if the AI tool was used by the physician as augmenting technology only. A very patient-centred usage of AI was examined in a qualitative study from the US. 13 patients tested an AI tool which helped them to interpret and understand their written X-ray reports, thus aiming to adequately meet their information needs. They rated the usage of AI in this special setting as positive, but however, in addition to the previously described perceived risk factors, complained about the missing empathy of the system. Another very patient-centric use case for AI is the application of symptom checkers which are available online. In an online survey among US American users of such symptom checker, more than 90% of the 329 respondents perceived it as useful diagnostic tool providing helpful information and in still half of the cases leading to positive health effects.

Research gap and objectives

While many previous publications assessed the perception of AI with respect to a specific technical application, the general perception on all different aspects of AI stays widely unclear. Thus, we aimed to collect data on the respondent’s current awareness about AI, which is present without supplying additional external knowledge. Examinations on the consumers’ perceptions of AI in medicine are frequently carried out using online survey tools[36,37] or are based on analyses of social media. These approaches offer a quick and simple generation of a big number of respondents, but it must not be overlooked that they are prone to methodological problems, like in particular, a high risk of a selection bias excluding persons without internet access either through technical deficiencies or through missing skills. The situation in Germany is of special interest, since the speed of digitization in health care system has been particularly slow in Germany. In 2018, Germany ranked on the second last rank among 17 countries in a study examining the extent of this digitization. A possible reason for that issue is the strict interpretation of data protection legislation in Germany compared to other countries.[41,42] Other obstacles are a fragmented health care system with a large number of stakeholders, partly lacking willingness and necessary organizational structures. It is therefore also of particular interest whether this situation also reflects in the perceptions and attitudes of patients as laypersons in this field, regardless of whether they are considered a cause or a consequence of delayed digitization. Therefore, the objective of this study was to investigate the perception of AI in healthcare using a paper-based questionnaire focussing on patients or their companions who were in direct contact with the healthcare system. Additionally, we aimed to examine whether the perception of AI in healthcare is affected by the digital affinity or sociodemographic factors of the respondents.

Materials and methods

We conducted a cross-sectional study using a paper-based questionnaire with patients and their companions at a tertiary referral hospital in Aachen, Germany. The reporting is carried out according to the Consensus-Based Checklist for Reporting of Survey Studies (CROSS).

Questionnaire development

To identify relevant subjects and items for the survey, we carried out a selective literature review analysing systematic reviews and original articles.[2,12,30,45-48] Additionally, open and explorative interviews were carried out with six healthcare professionals. The structure of the resulting questionnaire is given in Table 1.
Table 1.

Structure of the final resulting questionnaire.

SubsectionNumber of questions
Evaluation of the respondent’s technical affinity9 questions
Among them: Self-reported technical affinity score - 3 questions
Perception of different aspects of AI in healthcare26 questions
Subsections:
- ‘AI brings advantages for patients’ - 6 questions
- ‘Patients fear AI’ - 4 questions
- ‘Patients are worried about physicians’ low AI competence’ - 5 questions
- ‘AI needs to be controlled’ - 3 questions
Details on sociodemographic characteristics5 questions
Structure of the final resulting questionnaire. The questions in the section “Perception of different aspects of AI in healthcare” used both positive and negative wording. The perception and self-assessment questions were answered on a five-point Likert scale or yes/no questions. Finally, respondents had the opportunity to give feedback in free text form. Since the target population consisted of patients and their attendance waiting for an appointment in a walk-in clinic, it was emphasized that it is possible to complete the questionnaire within 15 min. The healthcare professionals checked the first draft of the questionnaire for clearness, comprehensibility and possible mistakable phrases. Their feedback was implemented immediately.

Pretesting

According to Perneger et al., a sample size of 30 respondents was chosen for the pre-testing of the survey. The questionnaire was pretested in a group of volunteers that was assumed to be as similar as possible to the expected final sample. For the sociodemographic characteristics of the pre-test population, please refer to the Appendix (Appendix 1, Table A1). The results of the pre-test were analysed for reliability using Cronbach’s alpha and inter-item correlation and corrected item-total correlation. The Cronbach’s alpha of the technical affinity rating in section 1 (“Technical affinity”) and the four subsections in section 2 (“Perception of AI in healthcare”) are given in the Appendix (Appendix 1, Tables A2 and A3). Questions with a corrected item-total correlation of <0.3 or an inter-item correlation of >0.8 were removed from the subsections to allow correct evaluation of the subsections. However, these questions were not removed from the questionnaire, due to their relevance. Additional improvement on clearness were made according to the feedback of the pre-test participants. The pre-test revealed that about 75% of the participants were able to complete the questionnaire within 15 min. The final questionnaire is provided as a translated, English version in the Appendix 2.

Sample characteristics and sample size

The participants were opportunistically recruited from the waiting area of the premedication outpatient clinics of the Department of Anaesthesia of the University Hospital RWTH Aachen, Germany. Most participants consisted of inpatients and outpatients, but also accompanying relatives or friends were invited to fill in the questionnaire. The study was carried out from December 2019 to February 2020. The inclusion criteria were (a) the capability to read and understand the information given in the questionnaire and (b) the willingness to take part in the survey. There were no explicit exclusion criteria. The results of our survey should be transferable to the German population as a whole. Given a population of 83 million inhabitants, a confidence level of 95% and an error margin of 5%, this resulted in an estimated sample size of 385 respondents. We added a safety margin of 20% resulting in a target sample size of 462 participants. The University Hospital in Aachen serves as a supramaximal healthcare provider covering the entire spectrum of medicine with 36 specialist clinics, 28 institutes and six interdisciplinary units. Due to its geographical position and its extended infrastructure, the hospital covers not only the city of Aachen, but also the surrounding midsize cities which are still affected by the structural change as well as some more rural areas. It covers all surgical procedures, from minor outpatient surgeries to procedures that last several hours and require subsequent intensive care. Therefore, the hospital serves a highly diverse spectrum of patients and is thus particularly suitable for the generation of a broad survey population.

Survey administration

The present study was reviewed and approved by the local Ethics Committee at the RWTH Aachen University Faculty of Medicine (EK 307/19). Since a scientific publication of the results in an anonymous manner was announced on the first page, the completion of the questionnaire was understood as consent and explicit written consent was waived. The survey was conducted in compliance with all relevant regulations including data privacy legislation. At their visit at the premedication outpatient clinics, patients were asked for willingness to answer the questionnaire during their waiting time for their appointment with the anaesthesiologist. A member of the research group handed out the paper-based questionnaire and answered possible questions. After completion, participants put the questionnaire in an opaque box to ensure privacy.

Statistical analysis

The completed questionnaires were digitalized manually before analysis. Descriptive statistics were used to characterize the sample by age, gender, highest educational attainment, type of current occupation and an occupation in the health sector. Insufficiently filled in questionnaires (more than one-third of missing questions or sections not completed) were not included into the analysis. The results were expected statistically significant if p < .05. Continuous variables were expressed as mean value and standard deviation. Categorical variables and Likert scale ratings were calculated both as mean value and standard deviation and in absolute numbers and proportion. Missing data were not imputed. For evaluation of a subsection, the mean values of the different answers were combined to one score for the respective subsection. To analyse the influence of biometric characteristics on self-reported technical affinity, we performed an analysis of covariance (ANCOVA). Age, gender, the level of education (combined as low, medium and high) and an occupation in healthcare were used as dependent variables. The null hypothesis that the expected value of self-reported technical affinity is identical in the different classes studied (gender, education and healthcare profession) adjusted for age was tested with an F-test at the 5% significance level. Additionally, we performed a multivariable logistic regression analysis to evaluate the influence of the previously mentioned sociodemographic factors and the self-reported technical affinity on the general perception on AI in healthcare. The null hypothesis that the odds ratio for all influencing variables is equal to one was tested with a Wald test at a significance level of 5%. The linearity in the logit was graphically assessed by plotting the observed values of age and technical affinity against the predicted logits of general perception on AI in healthcare. The possible presence of multicollinearity was discussed using a correlation matrix. All statistical analyses were carried out using SAS Version 9.4 (Cary, North Carolina, USA).

Results

Respondent characteristics

A total of 452 participants completed the questionnaire to the required extent and were included in the analysis; ten questionnaires were excluded. The demographic characteristics of the respondents are presented in Table 1. Age, gender and level of education showed a nearly equal distribution. More than half of the respondents were employed. About 25% had a medical or health professional background.

Usage of technical devices and self-estimated technical affinity

The usage of technical devices was high among the respondents (Figure 1). A total of 92.02% of the participants stated a daily use of computers, smartphones and other technical devices. As well, the internet was used on a daily basis by 84.39%. While the vast majority (96.2%) reported owning a smartphone, the usage of wearables for monitoring of body functions (22.54%) or of medical apps (34%) was more uncommon. In the self-rated technical affinity (Figure 1), the majority of respondents reported positive values for confidence using technical devices, preference for work with technical devices and competence for usage of new devices. More than 90% of the respondents stated, they had already read or heard about AI, but only 24% reported good or expert knowledge. The absolute majority said they knew roughly what AI was (Figure 2).
Figure 1.

Usage of information technology and self-reported technical affinity.

Figure 2.

Self-assessment of previous knowledge in relation to artificial intelligence.

Usage of information technology and self-reported technical affinity. Self-assessment of previous knowledge in relation to artificial intelligence.

Perception on AI in healthcare

The results of the survey regarding the perception of AI usage in healthcare can be found in Figure 3. Asked on their general perception, a clear majority favoured the use of AI in medicine and healthcare: 53.18% of the respondents rated it as positive or very positive. On the opposite, only 4.77% had a negative or very negative opinion. The remaining of the respondents gave a neutral or no estimation.
Figure 3.

Perception of different aspects of AI in healthcare.

Perception of different aspects of AI in healthcare. In contrast to this statement, the factor stating that AI brings benefits for patients received a slight agreement in the mean (2.17  ±  0.55) (see Figure 4). While a clear majority (55.73%) of respondents agreed with the statement that AI has benefits for patients, the proportion of respondents, who would ask for a usage of AI in their personal treatment, decreased to 41.2%. Nearly 20% would even refuse to be treated using AI-based applications. Furthermore, the agreement and disagreement to the statements that assume a relieving of work burden for the physicians by the use of AI are nearly equally distributed revealing respondents with a more pessimistic estimations beside the positive expectations.
Figure 4.

Boxplot representing the subsections characterized by four statements. The box corresponds to the interquartile range (IQR) with the median (line inside the box) and the whiskers representing 1.5 times the IQR. The mean is depicted by the diamond. Outlier are shown by single dots.

Boxplot representing the subsections characterized by four statements. The box corresponds to the interquartile range (IQR) with the median (line inside the box) and the whiskers representing 1.5 times the IQR. The mean is depicted by the diamond. Outlier are shown by single dots. General fear of AI in healthcare itself as well as fear of low AI competence in healthcare providers tended to be less important to the respondents (see Figure 4). However, the extent of disagreement is in a similar range as the former described factor. For the rating of the fear for AI itself, the respondents denied fearing the influence of AI on medical treatments. Consistently, they would also not claim to stop or prevent AI usage in medicine in general. The risk of a wrong decision made by a physician is obviously seen similar to an AI made mistake. However, there seem to be relevant safety concerns worrying about cyberattacks that could sabotage the AI systems. Among the respondents, there are predominantly positive reactions with respect to the role of physicians using AI. Thus, there is a bigger part of respondents who rate physicians sufficiently competent to handle the challenges of AI usage, like an impairment of the relationship between physician and patient. Regarding AI potentially hampering the development of the physician’s clinical abilities, the ratings are nearly equally distributed. The generally high confidence in physicians is also reflected in a stronger disagreement to the question if a bad prognostic value could affect the physician’s effort to save a patient, which is disagreed by 43.49% of the respondents. A very strong agreement was measured regarding the need for control when using AI in healthcare. Thus, the statement that a final decision on diagnosis or therapy should always be in the hands of a physicians received the highest degree of agreement (96.02%) in the whole survey. However, a very high agreement was measured for the need for a functional check of an AI-based application through an independent institution (76.46%) as well as a scientifically proven benefit before usage at bedside (73.37%). Other questions yielded some more unexpected answers. Although they belonged thematically to one of the factors already described, they were removed from the respective factor due to the calculations from the pre-test. From some questions, one can conclude that patients and their relatives have a very high opinion of their physicians and their competences. So, nearly two-thirds of the respondents do not think that physicians might be less meaningful in future medicine. An even higher proportion (73.72%) would trust a physician more than an AI-based system and 69.44% would want the physician to override an AI’s recommendation if he/she comes to different conclusions. The point that physicians might possibly not have enough knowledge about AI to use it at the bedside receives more disagreement than agreement. A last important issue for further development of data-driven approaches in medicine is the willingness of patients to make their health-related data available for non-commercial research purposes, which is agreed by 74.78% of the population. The participating patients and companions were asked to express their perceptions on AI in healthcare on a 5-point Likert scale enabling them to take a neutral position. It was striking that in many questions a very high rate of respondents chose this option, in two cases even more than 50%. In 13 questions, the neutral option was the most rated answer.

Free text statements

Forty-two respondents gave free text statements using the offered box at the end of the questionnaire. Many of them used this opportunity to explicitly stress their opinion on certain aspects, which were already contained in the questionnaire. The statements of 10 respondents were very similar covering a combination of three topics. They stated that human physicians have to keep the control over AI applications in healthcare. These applications must be used as a support tool but never perform any independent decisions. It was obviously important to these respondents to point out that AI must never replace a physician and that the ‘medical art’ must remain the basis for the treatment of patients. Another big group of respondents expressed positively about the usage of AI in healthcare. However, it was not always clear from the written statements whether the respondents were in favour of this development or whether they saw it as an unstoppable fact that could not be influenced anyway. Even if no explicit negative statements were made, a fatalistic interpretation of the statements could be assumed. A third group of respondents showed scepticism that AI would be helpful to reduce the workload of the physicians and nurses. Individual respondents shed light on yet other aspects: one respondent expressed concern about the unavailability of AI systems, for example, in the event of massive blackouts and the associated inability of physicians to act. Another respondent warned that every AI algorithm always can be as good as the underlying data bringing up the issue of data quality. Finally, another respondent expressed incomprehension that so much effort is being invested in AI and medical research, while other more important issues like an accelerating climate change remain insufficiently attended.

Influence of sociodemographic factors on technical affinity

To evaluate the influence of sociodemographic factors on the self-reported technical affinity, an ANCOVA was carried out. It could be shown that higher age (F(1/419)  =  91.55, p < .0001), female gender (F(1/419)  =  6.11, p  =  .0138) and lower educational levels (F(2/419)  =  7.88, p  =  .0004) were associated with a lower technical affinity. Regarding the level of education, the effect across the classes was that the higher the subjects were educated, the higher their affinity for technology. A former or current occupation in healthcare did not have an influence on technical affinity.

Influence of sociodemographic factors and technical affinity on the general perception of AI usage in healthcare

By use of a multivariable logistic regression model, we assessed the influence of the sociodemographic characteristics and the self-reported technical affinity on the general perception of AI usage in healthcare. The null hypothesis that all odds ratios are OR = 1 for all influence variables was rejected (Wald test  =  53.996, p < .0001). Gender, educational level and technical affinity showed a significant influence on the general perception of AI usage in healthcare. The ORs and 95% confidence intervals (CIs) of the examined variables are given in Table 2. The self-reported technical affinity showed the strongest relation with the general AI perception (see Figure 5).
Table 2.

Sociodemographic characteristics of the respondents. Data are given as mean  ±  SD or n (%).

N = 452
Age46.69  ±  16.03
 – <2012 (2.65)
 – 20–2970 (15.49)
 – 30–39102 (22.57)
 – 40–4955 (12.17)
 – 50–59102 (22.57)
 – 60–6973 (16.15)
 – 70–7933 (7.30)
 – >805 (1.11)
Gender
 – Male206 (45.68)
 – Female244 (54.10)
 – Non-binary1 (0.22)
Level of education
 – ‘Low’
  ○ No school leaving certificate2 (0.45)
  ○ Primary school (Volksschule)24 (5.37)
 – ‘Medium’
  ○ Secondary school (Hauptschule)55 (12.30)
  ○ Secondary school (Mittlere Reife)115 (25.73)
 – ‘High’
  ○ A-levels/technical baccalaureate113 (25.28)
  ○ (Technical) college/university130 (29.08)
 – Other8 (1.79)
Current occupation
 – Pupil4 (0.89)
 – Apprentice10 (2.22)
 – Student19 (4.22)
 – Househusband/wife30 (6.67)
 – Employee245 (54.44)
 – Self-employed26 (5.78)
 – Civil servant12 (2.67)
 – Job seeking6 (1.33)
 – Retirement87 (19.33)
 – Other11 (2.44)
Healthcare professional
 – Yes112 (25.40)
 – No329 (74.60)
Table 3.

Multivariate regression analysis, dependent variable: general perception of AI in healthcare. Wald Chi square and p-value from type 3 analysis of effects.

VariableWald chi squareOR95% CIp-value
Age0.6350.9940.9801.0090.4257
Female gender16.0612.2401.5103.324<0.0001
Educational level11.053       0.004
- Medium vs low1.0390.4612.339
- High vs low0.5140.2241.177
Occupation in healthcare0.2320.8960.5741.3990.6299
Technical affinity19.2340.5590.4320.725<0.0001
Figure 5.

Heatmap depicting the correlation between self-reported technical affinity and general perception on AI in healthcare.

Heatmap depicting the correlation between self-reported technical affinity and general perception on AI in healthcare. Sociodemographic characteristics of the respondents. Data are given as mean  ±  SD or n (%). Multivariate regression analysis, dependent variable: general perception of AI in healthcare. Wald Chi square and p-value from type 3 analysis of effects.

Discussion

Although the number of newly constructed AI-based applications for healthcare steadily increases and efforts to implement these models into clinical routine go on, surprisingly, the scientific interest on how patients perceive these developments came into the scientific focus just recently. However, this question is subject to many discussions among medical professionals and computer scientists as well as the interested public. Depending on personal beliefs, the positions in these discussions range from completely positive towards completely negative. Hence, it was the main objective to evaluate, where patients in general locate themselves within this continuous spectrum. Our results show that the majority of patients and their companions in Germany have a positive or very positive attitude to the usage of AI in healthcare, although their respective knowledge is moderate. However, the general perception seems to differ relevantly between certain groups of respondents. Elderly, female or less educated persons as well as those having a low technical affinity have a more sceptical view on AI in healthcare. Common to all respondents is the desire for intensive monitoring of AI by physicians and their rejection of too much autonomy of the systems. We consider these results relevant for all stakeholders, who are working on the different aspects of AI in medicine, like physicians, AI developers, healthcare industry, insurances and regulatory bodies. Our results can be helpful for the planning of strategies to guide the further development and implementation of AI in health care. Especially, an active involvement of patients into their treatment and their collaboration is a critical step to improve clinical outcomes. This applies not only if patients use AI themselves but also if they are treated under AI usage. Knowledge on special opinions and requirements of certain patient groups allow a specific adaption of developing and implementing measures.

Study population

We aimed to examine a group of persons that is in direct contact to the healthcare system that is representative for the German population. The biographic parameters of our study sample were similar to the total population. In comparison to the German population, our respondents were 2.2 years older in the mean and our cohort included more female than male respondents. This difference was about 3% higher in our group than in the total population. As far as age and gender ratio are concerned, it can be assumed that our sample covers the total German population with sufficient accuracy. The proportion of university graduates was nearly twice as high in our cohort, which might be due to the high number of academic and research centres in the district served by the hospital. The ratio of persons who work or have worked in the healthcare sector seems quite high (25.4%), since the proportion of health care workers who are actively employed is given as 12.2% for Germany as a whole. Even if our percentage includes retired health workers as well, their high proportion remains remarkable, and its reasons remain unclear in the end. Most respondents generally classified themselves as having a higher affinity for technology. However, it can be noted that this affinity does not easily reach the personal health sector. For example, while more than 96% of respondents use a smartphone, which is even higher than the German average, only about a third uses health-related apps on it. Nevertheless, the proportion of health-app users is slightly higher in our sample than reported for Germany before. Other authors came to similar results.[58-60]

General perception of AI usage in healthcare

Our survey showed a general open-mindedness of patients and their relatives towards the usage of AI in healthcare. More than 50% rated its usage as positive or very positive. A similar proportion expects benefits for patients through the usage of AI. However, this view of AI by patients, which seems very positive at first glance, should not tempt to take an uncritical view. There are several aspects that need to be considered to put the results into the right context. Firstly, a relevant part of respondents chose the neutral option of the Likert-scale. They may be truly neutral, they may not have had enough information to make an informed choice or they may have been trying to avoid socially undesirable responses. Although we took care to make the completion of the questionnaire as anonymous as possible, it cannot be ruled out that some respondents chose the most harmless option. A much more likely option is an insufficient level of information of the respondents, since in the self-rating, only one-third stated to be able to explain the meaning of AI. The big majority only knew ‘somehow’ what it meant or even less, which was in a similar range already shown for the German as well as other populations.[21,62] Much more explicitly, our data point out that an application of AI at the bedside is only acceptable for the vast majority of patients if they can be sure that the AI is under continuous control and more specifically, they expect this control to be taken over by a physician. Several respondents strengthened this point in their free text notes explicitly. This finding was already demonstrated in several previous studies.[8,22,23] Exceeding the extent of agreement in these studies that AI needs to be controlled, in our survey the agreement nearly reached unanimity (96.02%). At the latest when there is a discrepancy between a physician's and the AI's assessment, the positive rating of AI ends and patients rely on their human physicians. This attitude also includes the permission for a physician to override an AI recommendation. These points comply with the finding from a multitude of studies that patients trust human physicians much more than an AI system,[22-28] which could also be reproduced in our study as well. Accompanying to the supervision by a medical professional, additional preclinical measures like a scientific evaluation and an independent certification of AI systems are requested by patients.[8,10] Uncertainties among patients in this context could also be the reason why, despite a very positive attitude in general, only a smaller proportion of patients in our study would agree to be treated using AI personally and other patients would even refuse an AI treatment, although their proportion was much smaller than in previous studies. In order to reach this sceptical group, it would be urgently necessary to combine the clinical implementation of AI methods with intensive information campaigns in order to overcome obvious reservations in the best possible way. However, the great mistrust of AI in the healthcare sector expressed by several stakeholders in medicine but also public discussions could not be demonstrated in our survey. The most likely explanation for this discrepancy is the assumption of patients that they expect a physician to be the gatekeeper for AI decisions. This is reflected by the high disagreement to the statements that physicians would play a less important role in the future treatment. So, in concordance with experts from the field,[63-65] patients seem to prefer a cooperation of AI and human physicians, in which the physicians make use of their humanistic skills, like empathy, communication and shared decision making, while the AI offers its comprehensive knowledge and its fast and precise analysis of patient data. Although patients disagree that physicians would not have enough knowledge about AI to use it in their daily work, in contrast, previous research showed that they are frequently lacking a sufficient knowledge about AI and its principles.[14,63,64] In this context, and with the increasing introduction of AI into clinical practice, there should be a stronger focus on AI education, especially in the training of future physicians.[65,66]

Influencing factors on AI perception

Our survey showed that older patients, women and persons with lower education had a more cautious view on the healthcare-related usage of AI. Less surprising, the personal technical affinity had an even stronger effect on the perception of AI usage. Thus, the fact that our survey population had a higher education compared to the German total population could bias the general view towards a more positive perception. Nevertheless, these influencing factors are certainly not new and are consistent with several previous publications.[58,67-69] This divergent perception can lead to reduced usage of novel technologies in certain groups and is commonly described as ‘digital divide’ or ‘digital gap’. While many think of these key words in terms of global phenomena such as limited internet access in developing countries, many authors emphasize that sociodemographic factors within a developed society can influence access to novel digital technologies as well.[58,67,70,71] For example, it is conceivable that older or less educated people may disagree to a treatment under AI usage due to concerns primarily resulting from a lack of information. Therefore, these underrepresented groups need to be given special consideration when designing programmes that aim to provide knowledge about new digital technologies in healthcare. For instance, the European Union already addressed this issue and started a ‘Digital Inclusion’ Initiative within the ‘Shaping Europe's digital future’ programme.

Ethical dimensions

There is a central dilemma when applying AI in medical treatment: who is ethically accountable for a decision which arises from the cooperation of a physician and an AI tool ? Due to the ‘black box’ character of the most AI algorithms, it is difficult for physicians to understand how these algorithms create their recommendations. However, this understanding is necessary to impose an accountability on a medical professional. Similar to ‘analogous’ medicine, a physician should always be able to justify decisions and name the factors that led him or her there. If a physician makes clinical decisions in cooperation with a tool, whose decisions are unexplainable by design, it must be discussed whether such an opaque system should be used. Secondly, it is unclear who should be responsible and legally liable if a patient was harmed as a result of a clinician's usage of an AI tool. Interestingly, the respondents in our survey who gave a non-neutral response impressively represent this dilemma by splitting half and half into a group who sees the physician as responsible and a group who disagreed. Another relevant aspect getting increasing public attention is algorithmic fairness, representing the absence of biases in AI models to the best possible way. These biases can derive from the data included into the algorithm or from the algorithm itself. All AI algorithms rely on big training data sets for its development. If the training data population differs from the population, the algorithm is intended to be used on or a subpopulation is less represented in this training data, for example, due to their reduced contact to the health system, the ability of an AI model to make precise predictions on these patients is relevantly reduced. This so-called representation bias can be a relevant source of discrimination based on age, gender, skin colour, ethnic origin, financial or occupational status, and other factors. However, also factors within the model itself might cause bias and consecutive harm, if a model is used for a global population although there are subgroups that should be considered differently (aggregation bias), if two metrics of the algorithm (e.g. overall accuracy vs sensitivity) impair each other (learning bias), if the dataset to test an AI model differs from the intended population to use the AI on (evaluation bias) or if an AI model is actually used differently from the way in which it is intended (deployment bias). Patients are intended to benefit from the usage of AI in health care. However, it must never be lost sight of the fact that patients must bear the consequences first-hand in the event of a faulty or unethical implementation of a newly developed and implemented AI tool. So, since physicians and scientists impose the potential risks of their developments to other persons, there is a strong obligation to minimize these technical, ethical, and moral issues of AI to the lowest possible extent before a broad clinical implementation of AI.[8,77]

Limitations

We conducted a paper-based, single-centred survey about the general perception of AI usage in healthcare. With the aim of changing the existing opinion about AI as little as possible, we provided very little additional information about AI. Participants were only told that one part of AI in the field of medicine consists, for example, of extracting previously unknown correlations from large amounts of data to be able to diagnose diseases much earlier, to predict the effect of a certain therapy or to predict the course of disease. We are aware that this explanation already means a restriction of the huge field of AI, but we intended to prevent far-off or even unrealistic imaginations in the respondents, especially among those with lower AI knowledge. It is questionable, how meaningful a survey can be, if the participants are not familiar with the subject. However, it reflects a kind of real-world data and we have to realize that also this partially uninformed population is confronted with AI and forms opinions about it. Public discussions are held not only between experts or informed participants, but also less informed persons contribute. Thus, our aim was to capture the public opinion to be aware of possible reservations, which could influence the common perception. The location of the participating hospital is in a region with a strong technological and research focus. Even if a wide surrounding area is served, there is a certain probability that the positive aspects of AI are perceived more strongly in this environment than in a less technological environment. The data acquisition of our survey was finished before the world was struck by the COVID-19 pandemic, which caused a significant boost of digitalization of the health system as well as of the working and everyday life. Further studies will be necessary to address questions about the impact of the global pandemic on the perception of AI in healthcare.

Conclusion

Concluding, patients and their companions in Germany are open towards the usage of AI in healthcare. Although the knowledge about AI is only mediocre in general, a majority of respondents rates AI in healthcare as positive or very positive in general. It is noticeable that patients are more reluctant when AI usage concerns their own personal treatment. Older patients, women, persons with lower education and lower technical affinity had a more cautious view on the healthcare-related usage of AI. Notably, patients strengthen that it is essential that a physician controls the AI application and has the ultimate responsibility for diagnosis and therapy. Click here for additional data file. Supplemental material, sj-docx-1-dhj-10.1177_20552076221116772 for Attitudes and perception of artificial intelligence in healthcare: A cross-sectional survey among patients by Sebastian J Fritsch, Andrea Blankenheim, Alina Wahl, Petra Hetfeld, Oliver Maassen, Saskia Deffge, Julian Kunze, Rolf Rossaint, Morris Riedel, Gernot Marx and Johannes Bickenbach in Digital Health Click here for additional data file. Supplemental material, sj-docx-2-dhj-10.1177_20552076221116772 for Attitudes and perception of artificial intelligence in healthcare: A cross-sectional survey among patients by Sebastian J Fritsch, Andrea Blankenheim, Alina Wahl, Petra Hetfeld, Oliver Maassen, Saskia Deffge, Julian Kunze, Rolf Rossaint, Morris Riedel, Gernot Marx and Johannes Bickenbach in Digital Health
  57 in total

1.  Sample size for pre-tests of questionnaires.

Authors:  Thomas V Perneger; Delphine S Courvoisier; Patricia M Hudelson; Angèle Gayet-Ageron
Journal:  Qual Life Res       Date:  2014-07-10       Impact factor: 4.147

2.  Patients' perceptions of using artificial intelligence (AI)-based technology to comprehend radiology imaging data.

Authors:  Zhan Zhang; Daniel Citardi; Dakuo Wang; Yegin Genc; Juan Shan; Xiangmin Fan
Journal:  Health Informatics J       Date:  2021 Apr-Jun       Impact factor: 2.681

3.  Artificial intelligence in healthcare: opportunities and risk for future.

Authors:  Sri Sunarti; Ferry Fadzlul Rahman; Muhammad Naufal; Muhammad Risky; Kresna Febriyanto; Rusni Masnina
Journal:  Gac Sanit       Date:  2021       Impact factor: 2.139

4.  An Artificial Intelligence-based Mammography Screening Protocol for Breast Cancer: Outcome and Radiologist Workload.

Authors:  Andreas D Lauritzen; Alejandro Rodríguez-Ruiz; My Catarina von Euler-Chelpin; Elsebeth Lynge; Ilse Vejborg; Mads Nielsen; Nico Karssemeijer; Martin Lillholm
Journal:  Radiology       Date:  2022-04-19       Impact factor: 11.105

Review 5.  What the evidence shows about patient activation: better health outcomes and care experiences; fewer data on costs.

Authors:  Judith H Hibbard; Jessica Greene
Journal:  Health Aff (Millwood)       Date:  2013-02       Impact factor: 6.301

6.  Physician Confidence in Artificial Intelligence: An Online Mobile Survey.

Authors:  Songhee Oh; Jae Heon Kim; Sung-Woo Choi; Hee Jeong Lee; Jungrak Hong; Soon Hyo Kwon
Journal:  J Med Internet Res       Date:  2019-03-25       Impact factor: 5.428

7.  Patients' views of wearable devices and AI in healthcare: findings from the ComPaRe e-cohort.

Authors:  Viet-Thi Tran; Carolina Riveros; Philippe Ravaud
Journal:  NPJ Digit Med       Date:  2019-06-14

8.  Clinician and computer: a study on patient perceptions of artificial intelligence in skeletal radiography.

Authors:  Thomas York; Heloise Jenney; Gareth Jones
Journal:  BMJ Health Care Inform       Date:  2020-11

Review 9.  Exploring stakeholder attitudes towards AI in clinical practice.

Authors:  Ian A Scott; Stacy M Carter; Enrico Coiera
Journal:  BMJ Health Care Inform       Date:  2021-12

10.  Patient perceptions on data sharing and applying artificial intelligence to healthcare data: a cross sectional survey.

Authors:  Ravi Aggarwal; Soma Farag; Guy Martin; Hutan Ashrafian; Ara Darzi
Journal:  J Med Internet Res       Date:  2021-07-05       Impact factor: 5.428

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.