| Literature DB >> 34366701 |
Mohammad Naiseh1, Dena Al-Thani2, Nan Jiang1, Raian Ali2.
Abstract
Human-AI collaborative decision-making tools are being increasingly applied in critical domains such as healthcare. However, these tools are often seen as closed and intransparent for human decision-makers. An essential requirement for their success is the ability to provide explanations about themselves that are understandable and meaningful to the users. While explanations generally have positive connotations, studies showed that the assumption behind users interacting and engaging with these explanations could introduce trust calibration errors such as facilitating irrational or less thoughtful agreement or disagreement with the AI recommendation. In this paper, we explore how to help trust calibration through explanation interaction design. Our research method included two main phases. We first conducted a think-aloud study with 16 participants aiming to reveal main trust calibration errors concerning explainability in AI-Human collaborative decision-making tools. Then, we conducted two co-design sessions with eight participants to identify design principles and techniques for explanations that help trust calibration. As a conclusion of our research, we provide five design principles: Design for engagement, challenging habitual actions, attention guidance, friction and support training and learning. Our findings are meant to pave the way towards a more integrated framework for designing explanations with trust calibration as a primary goal.Entities:
Keywords: Explainable AI; Trust; Trust Calibration; User Centric AI
Year: 2021 PMID: 34366701 PMCID: PMC8327305 DOI: 10.1007/s11280-021-00916-0
Source DB: PubMed Journal: World Wide Web ISSN: 1386-145X Impact factor: 2.716
Figure 1Screening prescription classification AI-based system classification
Figure 2A sample of prescribing system interface supported with AI recommendations
Participants demographics for exploration and design phases
| Phase | Participant | Gender | Age | Role | Year of experience | Organisation |
|---|---|---|---|---|---|---|
| Exploration | P1 | Male | 20–30 | Medical Doctor | 1–5 | A |
| P2 | Male | 20–30 | Medical Doctor | 1–5 | A | |
| P3 | Male | 20–30 | Medical Doctor | 5–10 | B | |
| P4 | Female | 20–30 | Pharmacist | 5–10 | B | |
| P5 | Female | 20–30 | Pharmacist | 5–10 | C | |
| P6 | Male | 30–40 | Medical Doctor | 5–10 | A | |
| P7 | Male | 30–40 | Medical Doctor | 10–15 | A | |
| P8 | Male | 30–40 | Medical Doctor | 5–10 | B | |
| P9 | Female | 30–40 | Pharmacist | 10–15 | B | |
| P10 | Female | 30–40 | Pharmacist | 15–20 | C | |
| P11 | Female | 30–40 | Pharmacist | 10–15 | C | |
| P12 | Female | 30–40 | Pharmacist | 10–15 | C | |
| P13 | Female | 30–40 | Pharmacist | 10–15 | A | |
| P14 | Male | 40–50 | Medical Doctor | 15–20 | B | |
| P15 | Male | 40–50 | Medical Doctor | 15–20 | B | |
| P16 | Female | 40–50 | Pharmacist | 15–20 | C | |
| Co-design | P17 | Male | 20–30 | Medical Doctor | 1–5 | B |
| P18 | Male | 20–30 | Medical Doctor | 1–5 | B | |
| P19 | Female | 20–30 | Medical Doctor | 5–10 | D | |
| P20 | Female | 20–30 | Medical Doctor | 5–10 | B | |
| P21 | Female | 20–30 | Medical Doctor | 5–10 | D | |
| P22 | Male | 30–40 | Medical Doctor | 5–10 | B | |
| P23 | Female | 30–40 | Medical Doctor | 10–15 | C | |
| P24 | Male | 40–50 | Medical Doctor | 15–20 | B |
Figure 3Exploration phase workflow
Figure 4Co-design session workflow
The main types of errors made by participants when interacting with explanation interfaces
| Error type | Sub-category | Description |
|---|---|---|
| Skipping | Lack of curiosity | Participants had a lack of desire to know, learn or experience an explanation. They did not feel that the explanation motivated them to learn new ideas, resolve knowledge and assist them in the task |
| Perceived Goal impediment | Participants perceive the explanation as an interruption to their decision-making task and delayed their task goal | |
| Redundant information | Participants skipped the explanation because they felt it was repeating facts, and it did not add substantiations | |
| Perceived complexity | Participants skipped long explanations, as they did not want to engage in what they perceived to be an effortful experience | |
| Lack of domain context | Participants skipped explanations that were not reflective and contextualised to their domain knowledge context | |
| Misapplying | Misinterpretations | Participants misunderstood the explanations which lead to incorrect conclusions |
| Confirmatory search | Several participants searched for information that confirms their initial hypothesis, i.e., they were selective in what to hear and rely on | |
| Rush understanding | Participants incorrectly held a belief that they understand the AI deeper than they actually did, i.e., they failed to recognise the limits of their own understanding | |
| Habits formation | Participants became gradually less interested in the details of explanation and overlooked and perceived it to be familiar to them | |
| Mistrust | Participants felt that explanations were not reflecting knowledge, or they were suspicious, i.e., they voiced scepticism about the correctness and validity explanations |
The four main themes that emerged from the co-design phase
| Design technique | Definition |
|---|---|
| Abstraction | It refers to extracting and generating main features from the explanation and make it possible to present them at multiple abstractions and granularity levels |
| Control | It refers to providing customisation functionality to control the information presented in the explanation (e.g., grouping, ordering) |
| Cues | It refers to additional elements that can draw users’ attention and help guiding them in the process utilising the explanations |
| Adaptation | It refers to varying the explanations characteristics, e.g., information, abstraction level, cues, order and modalities, in response to an interaction context, i.e., the ability to communicate explanations differently in different settings |
Figure 5An example of two levels of abstraction design presented in our co-design study
Fig. 6An example of suggested control techniques designs for Local explanations in the co-design sessions
Figure 7Friction design example for calibrated trust goal