Why is it that people don’t trust Artificial Intelligence (AI)? The results of a recent survey puts the situation in a nutshell: (US-American) people trust AI to pick their films and music, but not much else. In contrast, they have no problems with old-fashioned automated systems because their understanding is that these simply do what they’re told. The question therefore is, how can we make AI systems more understandable and trustworthy?
Recently, I’ve had the privilege to be a temporary member of SAP’s AI Ethics team, led by SAP’s head of AI ethics, Sebastian Wieczorek. Being a psychologist and human-centricity practitioner, one topic specifically caught my attention: the aspect of AI transparency towards the individuals working with or affected by AI systems. This led me to the following questions:
Summarizing my learnings on the topic, I’m going to line out five steps that will help you create appropriate and meaningful AI understandability.
Explainability and interpretability are important principles for building trustworthy AI (see https://oecd.ai/en/ai-principles for example). The differentiation between these two concepts is not consistent across publications. The most useful definitions I found so far (Broniatowski, 2021) are providing a helpful perspective for human-centered product definition and design of AI systems:
AI explainability provides an answer to the question “What did the machine do to reach the output?”
It refers to the technological context of the AI, its mechanics, rules and algorithms, to the training data etc.
It is interesting for users who need to understand whether an AI is working properly, for example with regards to non-discrimination.
AI interpretability addresses the question: “What does the output mean for me and the problem I need a solution for?”
It refers to the respective domain / ‘real-life’ context of the AI application, for example, in case of a decision about a loan application.
I will use “understandable / understandability” as an umbrella term for explainability and interpretability whenever the concepts of explainability and interpretability as defined above are both being referred to.
Infographic 1: Definitions of AI understandability, AI explainability, and AI interpretability.
Let me now introduce four personas to help illustrate their different needs towards explainability / interpretability, depending on their task and role: The AI developer, the AI ethicist, the AI end-user, and the affected user. Expectations towards understandability are typically different for these roles. (on other taxonomies see for example Zytek, A. et al., 2022.) The level of professional expertise (technology and / or domain) of the understandability audience is an additional aspect worthwhile looking into.
Infographic 2: Different personas and their needs towards explainability and interpretability.
Next to the personas and their specifics, we also have the respective use cases with their characteristics and risks. After all, there is a difference whether we are talking about music suggestions or hiring processes.
Use cases from areas such as human resources, law enforcement, or the operation of critical infrastructure are strong candidates for a high-risk rating. Usually, you will find the respective definition in the ethical, regulatory and / or organizational frameworks that are relevant for you.
Implementing understandability features can address related requirements towards transparency, for example in the draft EU AI Act:
Infographic 3: Personas and their needs as regards high-risk use cases.
However, not every AI system needs to provide transparency in the sense of explainability or interpretability, for example, if the impact on financial, social, legal or other critical aspects is low or if the specific technology is very mature and regarded as safe and secure.
For implementing understandability that is suitable, appropriate, and usable for the respective roles and use cases, sometimes the output of technical tools can already be sufficient. For other contexts, however, that may not be the case.
Combining technical tools and a human-centered practice is a good approach to ensure relevance and usability of implemented understandability.
Depending on the respective use case, explainable AI (XAI) algorithms like SHAP or LIME can contribute significantly to the creation of AI understandability, for example, by providing:
Design practices can enhance AI understandability implementation in a variety of ways, for example by:
There are several resources available on how to utilize (and if necessary: adapt) design methodology to AI and specifically to the creation of explainable AI (see for example Liao et al., 2021).
Whichever combination of methods you choose for creating AI understandability features, the result needs to be as efficient, correct and useful as possible. To ensure the highest quality and relevance, consider technical criteria as applicable, including:
And of course human-centered criteria, for example:
Infographic 4: Human-centered and technical criteria for AI understandability.
Looking at understandability of AI systems from different angles – the technical perspective and a human-centered point of view – holds great potential for meaningful transparency to individuals working with or affected by AI systems. This approach helps to create the appropriate trust and best performance in human–AI collaboration.