Stakeholders in Explainable AI
Alun Preece, Dan Harborne (Cardiff), Dave Braines, Richard Tomsett (IBM UK), Supriyo Chakraborty (IBM US)
https://arxiv.org/abs/1810.00184
Stakeholders in Explainable AI Alun Preece, Dan Harborne (Cardiff), - - PowerPoint PPT Presentation
Stakeholders in Explainable AI Alun Preece, Dan Harborne (Cardiff), Dave Braines, Richard Tomsett (IBM UK), Supriyo Chakraborty (IBM US) https://arxiv.org/abs/1810.00184 Explainability in GOFAI Explainability in AI is not a new problem. In the
Alun Preece, Dan Harborne (Cardiff), Dave Braines, Richard Tomsett (IBM UK), Supriyo Chakraborty (IBM US)
https://arxiv.org/abs/1810.00184
Explainability in AI is not a new problem. In the last ‘AI summer’ (the expert systems boom, a.k.a., ‘Good Old Fashioned AI’) it was acknowledged that explainability was needed for
It was also realized that these require different forms of explanation, framed by developers’ vs end-users’ conceptual models. The problem wasn’t solved before funding dried up!
‘Interpretability’ is now preferred over ‘explainability’. For our purposes, an explanation is a message intended to convey the cause and reason for a system output; an interpretation is the understanding gained by the recipient of the message. We can still observe different motivations:
actually work
Again, different forms of explanation are required, due to the differing conceptual models of the recipients….
A useful distinction (Lipton 2017)
deep neural network model; decision tree; rule base)
workings ‘after the fact’ (e.g., visualization of salient features, explanation by similar training examples) Notes:
WHI workshop at ICML 2018 https://arxiv.org/abs/1806.07552
Developers are chiefly concerned with building AI applications. Theorists are chiefly concerned with understanding and advancing AI. Ethicists are chiefly concerned with fairness, accountability, transparency and related societal aspects of AI. Users are chiefly concerned with using AI systems. The first three of these communities are well-represented in the AI interpretability literature. The fourth will ultimately determine how long summer lasts.
Verification is about ‘building the system right’. Validation is about ‘building the right system’. [⚠Overgeneralization alert]
It is hard to envisage verification without (some) transparency. Post hoc explanations are valuable for validation.
Known knowns: things we train an AI system to know Known unknowns: things we train an AI system to predict Unknown knowns: things we know the AI system doesn’t know (i.e., things outside its bounds) Unknown unknowns: the key area of concern for all communities! Verification tends to be used to define the space of knowns. Validation is essential to define the space of unknowns.
Methods that derive at least in part from internal states of an AI system, e.g., deep Taylor decomposition, Google feature viz. Caveats:
non-axiomatic
ethicist communities
highlight meaningful features of the input, may make recipients less inclined to trust system…
Traffic congestion classifier: ‘congested’ image explanation by deep Taylor decomposition
Important to differentiate post hoc from transparent explanations, e.g., LIME local – approximations, to recipients. Explanation features favoured by subject-matter experts include
‘Congested’: explanation by example via influence functions
Composite explanation object that can be unpacked per a recipient’s requirements. Layer 1 — Traceability: transparency-based bindings to internal states of the system showing the system ‘did the thing right’ [main stakeholders: developers and theorists] Layer 2 — Justification: post-hoc semantic relationships between input and output features showing the system ‘did the right thing’ [main stakeholders: developers and users] Layer 3 — Assurance: post-hoc representations with explicit reference to policy/ontology to give confidence that the system ‘does the right thing’ [main stakeholders: users and ethicists]
Layer 1 (traceability): saliency map visualisation of input layer features for classification ‘gorilla’ Layer 2 (justification): ‘right for the right reasons’ semantic annotation of salient gorilla features Layer 3 (assurance): counterfactual examples showing that images
Explanation in AI is an old problem: ‘debuggability’ shouldn’t be conflated with ‘trustability’. We can identify four communities with different stakes in explanation: developers, theorists, ethicists users. Developers & theorists tend to focus on verification (knowns); users & ethicists are more interested in validation (unknowns). Can the needs of multiple stakeholders be addressed via ‘joined- up’, composite explanations? Despite a large and growing literature on explanation/interpretation in AI, the voice of users is under-represented. As in the 1980s, it will be the users that determine whether AI thrives!
This research was sponsored by the U.S. Army Research Laboratory and the UK Ministry of Defence under Agreement Number W911NF–16–3–0001. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the official policies, either expressed or implied, of the U.S. Army Research Laboratory, the U.S. Government, the UK Ministry of Defence or the UK Government. The U.S. and UK Governments are authorized to reproduce and distribute reprints for Government purposes notwithstanding any copy-right notation hereon.