| Literature DB >> 34398656 |
John A McDermid1, Yan Jia1, Zoe Porter1, Ibrahim Habli1.
Abstract
In recent years, several new technical methods have been developed to make AI-models more transparent and interpretable. These techniques are often referred to collectively as 'AI explainability' or 'XAI' methods. This paper presents an overview of XAI methods, and links them to stakeholder purposes for seeking an explanation. Because the underlying stakeholder purposes are broadly ethical in nature, we see this analysis as a contribution towards bringing together the technical and ethical dimensions of XAI. We emphasize that use of XAI methods must be linked to explanations of human decisions made during the development life cycle. Situated within that wider accountability framework, our analysis may offer a helpful starting point for designers, safety engineers, service providers and regulators who need to make practical judgements about which XAI methods to employ or to require. This article is part of the theme issue 'Towards symbiotic autonomous systems'.Entities:
Keywords: assurance; explainability; machine learning
Year: 2021 PMID: 34398656 PMCID: PMC8366909 DOI: 10.1098/rsta.2020.0363
Source DB: PubMed Journal: Philos Trans A Math Phys Eng Sci ISSN: 1364-503X Impact factor: 4.226
Figure 1Context and roles of explainability. (Online version in colour.)
Figure 2ROC Curves for the example NNs. (Online version in colour.)
Accuracy for the example NNs.
| accuracy | |
|---|---|
| CNN | 86.5% |
| DNN | 87.1% |
Figure 3Comparative feature importance. (a) CNN feature importance, (b) DNN feature importance. (Online version in colour.)
Illustrations of explainability requirements for different stakeholders and scenarios.
| dimension/example | regulation | investigation | service | service | decision support | decision support |
|---|---|---|---|---|---|---|
| stakeholders | regulator | accident investigatora | service provider | end user | expert user | prediction recipient |
| scenario | system approval | investigate accident or incident | system deployment | service use | decision support | decision support |
| purpose of explanation | confidence, compliance | clarity, compliance, continuous improvement | confidence, compliance, (continuous improvement) | challenge, consent and control | confidence, consent and control, challenge | challenge |
| timing of explanations | pre-deployment | post-incident | pre-deployment | same time as decision | same time as decision | same time as decision |
| data explainability | global | local, global | global | n.a. | local, global | local |
| model explainability | global (interpretable models, adversarial examples, influential instances) | global (permutation feature importance, counterfactual explanations, TreeSHAP) | global (interpretable models, adversarial examples, influential instances) | n.a. | global (permutation feature importance, interpretable models) | n.a. |
| prediction explainability | n.a. | local (KernelSHAP, counterfactual explanations) | n.a. | local (KernelSHAP, DeepLIFT, interpretable models) | local (interpretable models, counterfactual explanations) | local (KernelSHAP, DeepLIFT, interpretable models) |
aService Provider may investigate service ‘outages’ (incidents) and Lawyers/Courts may also investigate challenges from decision recipients, using similar methods.