Why, Who, What, When and How about Explainability in
Human-Agent Systems
JAAMAS Track
Avi Rosenfeld
Jerusalem College of Technology
Jerusalem
rosenfa@jct.ac.il
Ariella Richardson
Jerusalem College of Technology
Jerusalem
richards@jct.ac.il
ABSTRACT
This paper presents a survey of issues relating to explainability in
Human-Agent Systems. We consider fundamental questions about
the Why, Who, What, When and How of explainability. First, we
defne explainability and its relationship to the related terms of
interpretability, transparency, explicitness, and faithfulness. These
defnitions allow us to answer why explainability is needed in
the system, whom it is geared to and what explanations can be
generated to meet this need. We then consider when the user should
be presented with this information. Last, we consider how objective
and subjective measures can be used to evaluate the entire system.
This last question is the most encompassing as it needs to evaluate
all other issues regarding explainability.
KEYWORDS
Humanśagent systems ; XAI ; Machine learning interpretability ;
Machine learning transparency
ACM Reference Format:
Avi Rosenfeld and Ariella Richardson. 2020. Why, Who, What, When and
How about Explainability in Human-Agent Systems. In Proc. of the 19th
International Conference on Autonomous Agents and Multiagent Systems
(AAMAS 2020), B. An, N. Yorke-Smith, A. El Fallah Seghrouchni, G. Sukthankar
(eds.), Auckland, New Zealand, May 2020, IFAAMAS, 4 pages.
1 OVERVIEW
As the feld of Artifcial Intelligence matures and becomes ubiqui-
tous, there is a growing emergence of systems where people and
agents work together. These systems, often called Human-Agent
Systems or Human-Agent Cooperatives, have moved from theory
to reality in the many forms, including digital personal assistants,
recommendation systems, training and tutoring systems, service
robots, chat bots, planning systems and self-driving cars [2ś4, 7, 10ś
12, 14ś18, 20ś23, 25, 26, 28]. One key question surrounding these
systems is the type and quality of the information that must be
shared between the agents and the human-users during their inter-
actions.
We focus on one aspect of this human-agent interaction Ð the
internal level of explainability that agents using machine learning
must have regarding the decisions they make. Our overall goal is
to provide an extensive study of this issue in Human-Agent Sys-
tems. Towards this goal, our frst step is to formally and clearly
Proc. of the 19th International Conference on Autonomous Agents and Multiagent Systems
(AAMAS 2020), B. An, N. Yorke-Smith, A. El Fallah Seghrouchni, G. Sukthankar (eds.),
May 2020, Auckland, New Zealand. © 2020 International Foundation for Autonomous
Agents and Multiagent Systems (www.ifaamas.org). All rights reserved.
defne explainability, as well as the concepts of interpretability,
transparency, explicitness, and faithfulness that make a system
explainable. Through using these defnitions, we provide a clear
taxonomy regarding the Why, Who, What, When, and How about
explainability and stress the relationship of interpretability, trans-
parency, explicitness, and faithfulness to each of these issues.
This paper’s frst contribution is a clear defnition for explain-
ability and for the related terms: interpretability and transparency.
In defning these terms we also defne how explicitness and faith-
fulness are used within the context of Human-Agent Systems. A
summary of these defnitions is found in Table 1. In defning these
terms, we focus on the features and records that are used as training
input in the system, the supervised targets that need to be identifed,
and the machine learning algorithm used by the agent. We defne
L as the machine learning algorithm that is created from a set of
training records, . Each record ∈ contains values for a tuple
of ordered features, . Each feature is defned as ∈ . Thus, the
entire training set consists of × . While this model naturally
lends itself to tabular data, it can as easily be applied to other forms
of input such as texts, whereby are strings, or images whereby
are pixels. The objective of L is to properly ft × with regard to
the labeled targets ∈ .
To help visualize the relationship between explainability, in-
terpretability and transparency, please note Figure 1. Note that
interpretability includes six methods, including transparent models,
and also the non-transparent possibilities of model and outcome
tools, feature analysis, visualization methods, and prototype analy-
sis. Feature analysis can serve as a basis for creating transparent
models, on its own as a method of interpretability, or as a inter-
pretable component within model, outcome and visualization tools.
Similarly, visualization tools can help explain the entire model as a
global solution or as a localized interpretable element for specifc
outcomes of ∈ . Prototype analysis uses as the basis for in-
terpretability, and not , and can be used for visualization and/or
outcome analysis of ∈ . Interpretability is a means for providing
explainability, as per these terms’ defnitions in Table 1.
To date, many reasons have been suggested for making systems
explainable [1, 5, 6, 8, 9, 13, 24]: to justify its decisions so the human
participant can decide to accept them (provide control), to explain
the agent’s choices and guarantee safety concerns are met, to build
trust in the agent’s choices, especially if a mistake is suspected or
the human operator does not have experience with the system, to
explain the agent’s choices that ensure fair, ethical, and/or legal
decisions are made, to explain the agent’s choices and better eval-
uate or debug the system in previously unconsidered situations,
JAAMAS Track Paper AAMAS 2020, May 9–13, Auckland, New Zealand
2161