| Literature DB >> 35994331 |
Sebastian Isbanner1, Pauline O'Shaughnessy2, David Steel2, Scarlet Wilcock3, Stacy Carter4.
Abstract
BACKGROUND: Artificial intelligence (AI) for use in health care and social services is rapidly developing, but this has significant ethical, legal, and social implications. Theoretical and conceptual research in AI ethics needs to be complemented with empirical research to understand the values and judgments of members of the public, who will be the ultimate recipients of AI-enabled services.Entities:
Keywords: artificial intelligence; bioethics; consumer health informatics; social values; social welfare; surveys and questionnaires
Mesh:
Year: 2022 PMID: 35994331 PMCID: PMC9446139 DOI: 10.2196/37611
Source DB: PubMed Journal: J Med Internet Res ISSN: 1438-8871 Impact factor: 7.076
Summary of the variables collected in the Australian Values and Attitudes on Artificial Intelligence (AI) study.
| Type of variable | Question number and variable | Concepts tested |
| General support or opposition |
B01—how much do you support or oppose the development of AI in general (with multiple examples given)?a |
Broad support for or opposition to AI |
| Importance of different attributes of AI in health care scenarios |
C01—machine reads medical test, diagnoses, and recommends treatment C02—machine triages when you are unwell |
In relation to: C01—delegation of clinical decisions to an autonomous machine C02—automating decisions about need for health care services (time-sensitive) Importance of: Explanation Speed Accuracy Human contact Reducing system costs Fair treatment Responsibility |
| Importance of different attributes of AI in welfare scenarios |
D01—machine processes application for unemployment benefits (data sharing required) D02—chatbot advises about carer payments |
In relation to: D01—foregoing privacy as a barrier to access services D02—automation of information services Importance of: Explanation Speed Accuracy Human contact Reducing system costs Fair treatment Personal tailoring |
| Support for or opposition to AI in specific health care scenarios |
C03—nonexplainable hospital algorithms C04—data sharing for quality care C05—deskilling physicians |
C03—importance of explainable machine recommendations C04—importance of privacy (balanced against quality of care) C05—importance of retaining human clinical skills |
| Support for or opposition to AI in specific welfare scenarios |
D03—targeted compliance checking D04—nonexplainable job services D05—automated assignment of parent support with limited contestability |
D03—algorithmic targeting of punitive policy D04—importance of explainable machine recommendations D05—importance of contestability (balanced against accuracy) |
| Speed—human contact |
E01—trade-off between quicker, more convenient, more accurate health care and social services and more human contact and discretion in health care and social services |
E01—speed and convenience and accuracy vs human contact and discretion |
| Sociodemographic |
Age, gender, concession card type, and employment status; household income, education, household type, language other than English spoken at home, and general health Centrelink payment, employment field, relevant experience, relevant degree, life satisfaction, and disability |
Descriptive variables collected using standard sociodemographic questions |
| Geographic |
State or territory, capital city or rest of state, and SEIFAb (geographic measure of disadvantage) |
Descriptive variables collected using standard questions about location of residence |
| Lifestyle |
How often they check the internet, how often they post comments or images to social media, how often they post on blogs, forums, or interest groups, early adopter by type, and television viewing by type of viewing |
Variables collected for weighting purposes |
aVariables in italics were collected from both the Life in Australia and web-based panel samples; all others were collected from the web-based panel alone.
bSEIFA: Socio-Economic Indexes for Areas.
Sociodemographic composition of Australian artificial intelligence survey sample (weighted data only).
|
| Combined sample (n=4448), n (%) | Web-based panel (n=1950), n (%) | |
|
| |||
|
| Capital city | 2957 (66.48) | 1300 (66.67) |
|
| Rest of state | 1481 (33.3) | 640 (32.82) |
|
| Not stated or unknown | 10 (0.22) | 10 (0.51) |
|
| |||
|
| 18 to 34 | 1386 (31.16) | 637 (32.67) |
|
| 35 to 54 | 1472 (33.09) | 660 (33.85) |
|
| 55 to 74 | 1166 (26.21) | 497 (25.49) |
|
| ≥75 | 394 (8.86) | 156 (8) |
|
| Not stated or unknown | 30 (0.67) | 0 (0) |
|
| |||
|
| Men | 2180 (49.01) | 939 (48.15) |
|
| Women | 2259 (50.79) | 1011 (51.85) |
|
| Other | 9 (0.2) | 1 (0.05) |
|
| Not stated or unknown | 0 (0) | 0 (0) |
|
| |||
|
| Employed | 2709 (60.9) | 1061 (54.41) |
|
| Not employed | 1735 (39.01) | 890 (45.64) |
|
| Not stated or unknown | 4 (0.09) | 0 (0) |
|
| |||
|
| Postgraduate qualification | 529 (11.89) | 246 (12.62) |
|
| Undergraduate or diploma | 1393 (31.32) | 676 (34.67) |
|
| Vocational qualification | 937 (21.07) | 398 (20.41) |
|
| School qualification | 1492 (33.54) | 626 (32.1) |
|
| Not stated or unknown | 96 (2.16) | 5 (0.26) |
|
| |||
|
| ≥Aus $3000 (US $2086.20) | 635 (14.28) | 211 (10.82) |
|
| Aus $1500 to Aus $2999 (US $1043.10 to US $2085.50) | 1281 (28.8) | 589 (30.21) |
|
| Aus $500 to Aus $1499 (US $347.70 to US $1042.40) | 1646 (37.01) | 793 (40.67) |
|
| <Aus $500 (US $347.70) | 550 (12.37) | 261 (13.38) |
|
| None | 139 (3.13) | 70 (3.59) |
|
| Negative income | 34 (0.76) | 26 (1.33) |
|
| Not stated or unknown | 162 (3.64) | 0 (0) |
|
| |||
|
| Yes | 1036 (23.29) | 438 (22.46) |
|
| No | 3411 (76.69) | 1513 (77.59) |
|
| Not stated or unknown | 1 (0.02) | 0 (0) |
|
| |||
|
| Excellent | 549 (12.34) | 236 (12.1) |
|
| Very good | 1887 (42.42) | 837 (42.92) |
|
| Good | 1302 (29.27) | 562 (28.82) |
|
| Fair | 573 (12.88) | 255 (13.08) |
|
| Poor | 131 (2.95) | 59 (3.03) |
|
| Not stated or unknown | 6 (0.13) | 0 (0) |
Figure 1Responses to question B01: How much do you support or oppose the development of artificial intelligence?
Estimated percentages, mean, and 95% CIs for responses to question B01: How much do you support or oppose the development of artificial intelligence?a,b
|
| Estimated percentage (95% CI) | Design effect |
| Strongly support | 19.5 (17.9-21.1) | 1.87 |
| Somewhat support | 40.8 (38.9-42.8) | 1.84 |
| Neither support nor oppose | 21.9 (20.3-23.5) | 1.74 |
| Somewhat oppose | 9.2 (8.1-10.4) | 1.87 |
| Strongly oppose | 4.2 (3.5-5.1) | 1.76 |
| I don’t know | 4.4 (3.6-5.3) | 1.96 |
aPercentages and CIs adjusted for weighting.
bThe mean score was 2.35 (95% CI 2.31-2.39) with a design effect of 1.83.
Percentage of those who strongly support or somewhat support the development of artificial intelligence, 95% CIs, and P values for testing against 50%a.
|
| Categories deleted | |
|
| “Don’t know” | “Don’t know and neutral” |
| Estimated percentage support (95% CI) | 63.1 (61.1-65) | 81.8 (80-83.5) |
| <.001 | <.001 | |
| Design effect | 1.80 | 1.83 |
aPercentages and CIs adjusted for weighting.
bP value for adjusted Pearson F test for equal proportions in support and oppose categories.
Figure 2Responses to questions C03 to C05 and D03 to D05: support for or opposition to specific scenarios. AI: artificial intelligence.
Percentage of those supporting artificial intelligence in specific scenarios, 95% CIs, and P values for testing against 50%a.
| Domain and scenario | Estimated percentage in “support” or “strongly support” categories (95% CI) | Design effect | ||||||
|
| ||||||||
|
| Data sharing for quality care (C04c) | 42.3 (39.6-45.1) | <.001 | 1.62 | ||||
|
| Nonexplainable hospital algorithms (C03) | 29.1 (26.7-31.6) | <.001 | 1.57 | ||||
|
| Deskilling physicians (C05) | 27 (24.6-29.5) | <.001 | 1.57 | ||||
|
|
| |||||||
|
| Targeted compliance checking (D03) | 38.9 (36.2-41.7) | <.001 | 1.61 | ||||
|
| Automated parent support (contestability; D05) | 34.9 (32.3-37.6) | <.001 | 1.59 | ||||
|
| Nonexplainable job services (D04) | 31.2 (28.7-33.8) | <.001 | 1.56 | ||||
aPercentages and CIs adjusted for weighting.
bP value for adjusted Pearson F test for 50% proportions in categories 1 and 2 combined.
cCode in parentheses (eg, C04) indicates question number in instrument.
Proportion of respondents supporting artificial intelligence in specific scenarios, associated 95% CIs, and P values for testing against 50%; neutral responses deleteda.
| Domain and scenario | Estimated percentage in “support” or “strongly support” categories | Design effect | |||
|
| |||||
|
| Data sharing for quality care (C04c) | 57.8 (54.5-61.1) | <.001 | 1.63 | |
|
| Nonexplainable hospital algorithms (C03) | 41.1 (38-44.4) | <.001 | 1.58 | |
|
| Deskilling physicians (C05) | 35.8 (32.8-38.9) | <.001 | 1.58 | |
|
|
| ||||
|
| Targeted compliance checking (D03) | 54.1 (50.9-57.4) | .01 | 1.58 | |
|
| Automated parent support (contestability; D05) | 50.4 (47-53.7) | .82 | 1.62 | |
|
| Nonexplainable job services (D04) | 44.1 (40.8-47.4) | <.001 | 1.59 | |
aPercentages and CIs adjusted for weighting.
bP value for adjusted Pearson F test for 50% proportions in categories 1 and 2 combined.
cCode in parentheses (eg, C04) indicates question number in instrument.
Analysis of mean support for use of artificial intelligence (AI) in specific scenarios, 95% CIs, and P values for testing against a mean of 3. A score <3 represents support, and a score of >3 represents oppositiona.
| Domain and scenario | Estimated mean (95% CI) | Design effect | |||||
| General—support for the development of AI (B01c) | 2.35 (2.31-2.39) | <.001 | 1.83 | ||||
|
| |||||||
|
| Data sharing for quality care (C04) | 2.90 (2.83-2.98) | .01 | 1.65 | |||
|
| Nonexplainable hospital algorithms (C03) | 3.25 (3.18-3.32) | <.001 | 1.57 | |||
|
| Deskilling physicians (C05) | 3.39 (3.31-3.46) | <.001 | 1.62 | |||
|
| |||||||
|
| Targeted compliance checking (D03) | 2.98 (2.91-3.06) | .64 | 1.62 | |||
|
| Automated parent support (contestability; D05) | 3.06 (2.99-3.13) | .10 | 1.60 | |||
|
| Nonexplainable job services (D04) | 3.19 (3.12-3.26) | <.001 | 1.59 | |||
aMeans and CIs adjusted for weighting.
bP value for t test that the mean score was 3.0 using complex samples.
cCode in parentheses (eg, B01) indicates question number in instrument.
Estimated percentage of those who changed their response between the general question on the development of artificial intelligence and the specific scenarios and, of those who changed, the percentage that had a more negative attitude in the specific scenarios, with 95% CIs and the P value for the test of equal change in each directiona.
| Domain and scenario | Percentage of those who changed | Percentage of those who changed becoming more negative (95% CI) | Design effect | ||||||
|
| |||||||||
|
| Data sharing for quality care (C04c) | 60.2 | 70.8 (67.3-74) | <.001 | 1.59 | ||||
|
| Nonexplainable hospital algorithms (C03) | 65.6 | 81.4 (78.6-83.9) | <.001 | 1.53 | ||||
|
| Deskilling physicians (C05) | 70.6 | 83 (80.3-85.3) | <.001 | 1.56 | ||||
|
| |||||||||
|
| Targeted compliance checking (D03) | 63.8 | 71.9 (68.5-75) | <.001 | 1.65 | ||||
|
| Automated parent support (contestability; D05) | 65 | 76.1 (73-78.9) | <.001 | 1.56 | ||||
|
| Nonexplainable job services (D04) | 66.6 | 80.3 (77.5-82.9) | <.001 | 1.50 | ||||
aPercentages and CIs adjusted for weighting.
bAdjusted Pearson F test for equal proportions changing in each direction.
cCode in parentheses (eg, C04) indicates question number in instrument.
Estimated proportion of those who changed their response between 2 scenarios and, of those who changed, the percentage that expressed a more negative attitude in the second question, with 95% CIs and the P value for the test of equal change in each directiona.
| Domain and scenarios compared | Percentage of those who changed | Percentage of those who changed becoming more negative (95% CI) | Design effect | ||
|
| |||||
|
| C03c (explainability) vs C04d (data sharing) | 38.1 | 26.7 (22.7-31.1) | <.001 | 1.77 |
|
| C03 (explainability) vs C05e (deskilling) | 43.6 | 59.2 (55-63.3) | <.001 | 1.62 |
|
| C04 (data sharing) vs C05 (deskilling) | 45.7 | 77.9 (74.2-81.2) | <.001 | 1.69 |
|
| |||||
|
| D03f (compliance checking) vs D04g (explainability) | 41.7 | 64.2 (60-68.2) | <.001 | 1.60 |
|
| D03 (compliance checking) vs D05h (contestability) | 45.1 | 55.6 (51.4-59.6) | .008 | 1.59 |
|
| D04 (explainability) vs D05 (contestability) | 42.3 | 41.7 (37.6-45.9) | <.001 | 1.59 |
| Explainability in health vs in welfare—C03 vs D04 | 45.7 | 46.1 (42-50.2) | .06 | 1.64 | |
aPercentages and CIs adjusted for weighting.
bAdjusted Pearson F test for equal proportions changing in each direction.
cC03: nonexplainable hospital algorithms.
dC04: data sharing for quality care.
eC05: deskilling physicians.
fD03: targeted compliance checking.
gD04: nonexplainable job services.
hD05: automated parent support (contestability).
Figure 3Responses to questions C01 to C02 versus D01 to D02: summary and comparison of health (C) and welfare (D) scenarios. Numerical estimates <10% are not given.
Means, 95% CIs, and design effects for importance of values.
|
| Estimate of the meana (95% CI) | Design effect | ||
|
|
| |||
|
| Accuracy | 1.49 (1.46-1.53) | 1.98 | |
|
| Human contact | 1.78 (1.74-1.81) | 1.95 | |
|
| Responsibility | 1.78 (1.75-1.82) | 1.98 | |
|
| Explanation | 1.86 (1.82-1.90) | 1.96 | |
|
| Fairness | 1.87 (1.83-1.91) | 1.91 | |
|
| Speed | 2.08 (2.04-2.12) | 1.88 | |
|
| Reducing costs | 2.30 (2.25-2.34) | 1.92 | |
|
|
| |||
|
| Accuracy | 1.56 (1.51-1.61) | 1.73 | |
|
| Responsibility | 1.76 (1.71-1.81) | 1.75 | |
|
| Human contact | 1.81 (1.75-1.86) | 1.72 | |
|
| Explanation | 1.87 (1.82-1.93) | 1.76 | |
|
| Speed | 1.90 (1.85-1.95) | 1.64 | |
|
| Fairness | 1.94 (1.88-2.00) | 1.81 | |
|
| Reducing costs | 2.43 (2.36-2.50) | 1.74 | |
|
|
| |||
|
| Accuracy | 1.61 (1.56-1.65) | 1.53 | |
|
| Fairness | 1.80 (1.75-1.85) | 1.56 | |
|
| Explanation | 1.86 (1.80-1.91) | 1.61 | |
|
| Personal tailoring | 1.87 (1.82-1.92) | 1.58 | |
|
| Human contact | 1.88 (1.82-1.93) | 1.54 | |
|
| Speed | 1.99 (1.93-2.04) | 1.58 | |
|
| Reducing costs | 2.51 (2.45-2.58) | 1.59 | |
|
|
| |||
|
| Accuracy | 1.60 (1.55-1.64) | 1.6 | |
|
| Fairness | 1.81 (1.76-1.87) | 1.68 | |
|
| Personal tailoring | 1.82 (1.77-1.87) | 1.67 | |
|
| Human contact | 1.83 (1.77-1.88) | 1.63 | |
|
| Speed | 1.91 (1.86-1.97) | 1.71 | |
|
| Explanation | 2.02 (1.96-2.08) | 1.72 | |
|
| Reducing costs | 2.60 (2.54-2.67) | 1.71 | |
aMeans and CIs adjusted for weighting.
bCode (eg, C01) indicates question number in instrument.
Differences in mean responses on importance of attributes between 2 scenariosa.
| Domain and attribute | Mean difference (95% CI) | Design effect | |||||
|
| |||||||
|
| Explanation | −0.001 (−0.048 to 0.046) | .96 | 1.89 | |||
|
| Speed | 0.082 (0.040 to 0.123) | <.001 | 1.51 | |||
|
| Accuracy | −0.009 (−0.052 to 0.033) | .67 | 1.91 | |||
|
| Human contact | −0.012 (−0.060 to 0.036) | .63 | 2.12 | |||
|
| Responsibility | 0.007 (−0.035 to 0.050) | .73 | 1.88 | |||
|
| Reducing costs | −0.111 (−0.162 to −0.060) | <.001 | 1.99 | |||
|
| Fairness | −0.035 (−0.081 to 0.011) | .13 | 1.93 | |||
|
| |||||||
|
| Explanation | −0.164 (−0.215 to −0.113) | <.001 | 1.64 | |||
|
| Speed | 0.070 (0.029 to 0.111) | <.001 | 1.59 | |||
|
| Accuracy | 0.012 (−0.023 to 0.048) | .50 | 1.42 | |||
|
| Human contact | 0.049 (0.009 to 0.089) | .02 | 1.48 | |||
|
| Personal tailoring | 0.048 (0.006 to 0.090) | .02 | 1.58 | |||
|
| Reducing costs | −0.091 (−0.136 to −0.046) | <.001 | 1.54 | |||
|
| Fairness | −0.018 (−0.059 to 0.029) | .38 | 1.72 | |||
aMeans and CIs adjusted for weighting.
bP value for t test that the mean difference was 0 using complex samples.
cC01: machine reads medical test, diagnoses, and recommends treatment.
dC02: machine triages when you are unwell.
eD01: machine processes application for unemployment benefits (data sharing required).
fD02: chatbot advises about carer payments.
Estimated percentages of those who changed their responses on importance of values between 2 scenarios and, of those, the percentage that ranked the value to be more important in the first question than in the second question (C01 vs C02 or D01 vs D02), with associated 95% CIs and the P value for the test of equal cell proportionsa.
| Domain and values | Percentage of those who changed | Percentage ranking the value as more important in C01 (vs C02) or D01 (vs D02) (95% CI) | Design effect | ||||||
|
| |||||||||
|
| Explanation | 34.3 | 47.6 (42.8-52.4) | .33 | 1.68 | ||||
|
| Speed | 34.9 | 39.5 (35.2-44.1) | <.001 | 1.52 | ||||
|
| Accuracy | 25.1 | 49.5 (43.8-55.2) | .86 | 1.68 | ||||
|
| Human contact | 29.9 | 50.3 (45-55.5) | .92 | 1.70 | ||||
|
| Responsibility | 28.3 | 47.7 (42.5-53) | .40 | 1.69 | ||||
|
| Reducing costs | 33 | 59.2 (54.3-63.9) | <.001 | 1.66 | ||||
|
| Fairness | 29.3 | 53.7 (48.5-58.8) | .16 | 1.66 | ||||
|
| |||||||||
|
| Explanation | 39.6 | 63.7 (59.4-67.7) | <.001 | 1.55 | ||||
|
| Speed | 32.7 | 41.8 (37-46.6) | .001 | 1.66 | ||||
|
| Accuracy | 26.4 | 48.4 (43.2-53.7) | .56 | 1.57 | ||||
|
| Human contact | 30.7 | 43.9 (39.1-48.8) | .02 | 1.64 | ||||
|
| Personal tailoring | 33.1 | 43.9 (39.1-48.8) | .01 | 1.69 | ||||
|
| Reducing costs | 35.1 | 58.8 (54.3-63.1) | <.001 | 1.58 | ||||
|
| Fairness | 27.1 | 51.7 (46.3-57.1) | .53 | 1.70 | ||||
aPercentages and CIs adjusted for weighting.
bAdjusted Pearson F test for equal proportions.
cC01: machine reads medical test, diagnoses, and recommends treatment.
dC02: machine triages when you are unwell.
eD01: machine processes application for unemployment benefits (data sharing required).
fD02: chatbot advises about carer payments.
Figure 4Responses to question E01: speed, accuracy, and convenience versus human contact and discretion.
Speed, accuracy, and convenience versus human contact and discretion; estimated percentages; and 95% CIs for responses to question E01a.
|
| Estimate (95% CI) |
| 1: speed, convenience, and accuracy | 7.6 (6.2-9.1) |
| 2 | 12.7 (11-14.7) |
| 3 | 27.7 (25.3-30.3) |
| 4 | 28.5 (26.1-31.1) |
| 5: human contact and discretion | 23.5 (21.2-26) |
| Mean scoreb | 3.38 (3.41-3.54) |
aPercentages and CIs adjusted for weighting.
bP<.001 for testing that the mean score was 3; design effect=1.602.
Comparison of findings from the studies by Zhang and Dafoe [26] and the Monash Data Futures Institute [27] and from the Australian Values and Attitudes on Artificial Intelligence (AVA-AI): How much do you support or oppose the development of artificial intelligence?
|
| Zhang and Dafoe [ | Monash Data Futures Institute [ | AVA-AI (2020), weighted % |
| Strongly or somewhat support | 40.94 | 62.4 | 60.3 |
| Neither support nor oppose | 27.84 | 23 | 21.9 |
| Strongly or somewhat oppose | 21.69 | 10.5 | 13.4 |
| I don’t know | 9.54 | 4.1 | 4.4 |