Explainable Deep Reinforcement Learning: State of the Art and Challenges
Abstract.
Interpretability, explainability and transparency are key issues to introducing Artificial Intelligence methods in many critical domains: This is important due to ethical concerns and trust issues strongly connected to reliability, robustness, auditability and fairness, and has important consequences towards keeping the human in the loop in high levels of automation, especially in critical cases for decision making, where both (human and the machine) play important roles. While the research community has given much attention to explainability of closed (or black) prediction boxes, there are tremendous needs for explainability of closed-box methods that support agents to act autonomously in the real world. Reinforcement learning methods, and especially their deep versions, are such closed-box methods. In this article we aim to provide a review of state of the art methods for explainable deep reinforcement learning methods, taking also into account the needs of human operators - i.e., of those that take the actual and critical decisions in solving real-world problems. We provide a formal specification of the deep reinforcement learning explainability problems, and we identify the necessary components of a general explainable reinforcement learning framework. Based on these, we provide a comprehensive review of state of the art methods, categorizing them in classes according to the paradigm they follow, the interpretable models they use, and the surface representation of explanations provided. The article concludes identifying open questions and important challenges.
1. Introduction
EU ethical guidelines (EUEthicalGuidelines, ) on accountability, technical robustness and safety, oversight, privacy and data governance, non-discrimination and fairness, transparency, societal and environmental well-being specify the essential aspects to be considered by Artificial Intelligence (AI) and Machine Learning (ML) products and appliances. Concise analysis for accredited adherence of systems to EU ethical principles is an essential prerequisite for the acceptability of deploying and using any such application. In addition, the U.S.A National Security Commission reports in (AiCommission, ) that “Ethical considerations are an important facet of R&D, application, training, protection, and cooperation in AI. … Developing trustworthy AI systems is essential for operational integrity and adoption. It is closely connected to, and depends on, reliability, robustness, auditability, explainability and fairness. From the earliest phase, systems should be designed with ethics in mind”.
Designing and implementing systems with “deep knowledge” prevails in AI: We may recall the need of incorporating “deep knowledge” about the domains and problems addressed (in contrast to “shallow knowledge”) in expert and knowledge-based systems, mitigating brittleness, vulnerability to unforeseen circumstances, lack of ability to reason from “first principles” when needed, and inability to provide informative explanations on the reasoning towards reaching conclusions and prescribing solutions. Deep knowledge is indeed essential for AI systems, nowadays with some very emergent, important and tremendously larger/broader challenges, and with more severe, challenging consequences and inescapable needs: In contrast to interpretable systems that exploit explicit knowledge and whose reasoning mechanisms principles and inner-workings are more transparent to humans, today deep knowledge is encoded in hidden layers of multi-layered deep neural networks, in the form of weights on “synapses”, or in models that fit a value function ranging into sets of large number of actions in millions of states. Reaching this point of technological advances allows us to deploy autonomous systems acting in the real-world with significantly more advanced levels of perception, decision making and automation, all integrated in a single loop. These systems have not been instructed by direct advice, but have been trained on massive data sampled from the real world. Most importantly, the interpretability of knowledge and reasoning mechanisms has been lost, not only for the domain experts, or for the humans using these systems, but also for AI systems’ engineers and developers. This has huge impact on ethical concerns when deploying and using these systems, thus on systems’ accountability, liability, safety and trustworthiness.
The consequences are crucial especially when we aim at higher levels of automation, where models aim to support autonomous agents to act autonomously. The problem is addressed in the literature from different perspectives, depending on the targeted use of explanations: Machine learning experts and system developers need explanations of algorithms’ inner workings, while domain experts and operators using/working with autonomous systems need to understand how the system reaches specific decisions, especially when the solutions proposed are not intuitive or obvious, understanding models’ “whims” (to use an anthropocentric metaphor), objectives and capabilities.
Towards developing trustworthy and accountable autonomous systems (agents) to support automation in safety-critical real-world settings, this article aims to review state of the art methods regarding interpretability, explainability and transparency of reinforcement learning (RL) methods, with emphasis on their deep (DRL) versions: Currently we lack a concise understanding of the qualities of explainable and transparent DRL methods and we have to close this gap. This is important in many real-world complex settings in which DRL agents are deployed, spanning from (co-)driving autonomous vehicles, to supporting automation in air traffic management and other cases with time-critical decisions. The challenges to make DRL models interpretable/explainable, are great: Some of these are “inherited” from ML models in general, but for DRL methods some additional subtle and important issues need to be considered, given that DRL models exploit intertwined closed prediction box models and they do solve sequential decision tasks, aiming to increase the expected feature reward w.r.t. agents’ preferences, intended objectives, and constraints that the environment imposes (e.g. partial observability, delayed rewards, dynamics, multiple co-existing agents, etc):
Agents are expected to provide informative explanations even in cases where humans cannot explain their own way of thinking / acting on solving sequential decision problems, comprehensively and with sufficient clarity: This is for instance true in cases where it is easier for humans to demonstrate how they apply their knowledge and skills in example cases. For instance, when systems follow a data-driven approach learning from demonstrations, it may be hard to explicate the acquired knowledge. In these cases, providing concise “mainstream” explanations (e.g., explicating their knowledge and reasoning chains for expected rewards in long-term time horizons) may simply fail: Systems must be empowered with sophisticated means that allow them to demonstrate reasoning and problem solving abilities more intelligently (i.e., taking into account informative factors, providing examples of behaviour, or critical aspects of decision making), supporting users to understand systems’ rationale at various levels of detail. Let me provide an example on this. We may all have used the ”shoe laces” example in our early AI lectures to show how difficult it is for experts to express procedural knowledge on specific sequential tasks and problem solving, and thus, how difficult it is for systems to represent in symbolic and explicit ways such knowledge. However, nowadays, providing paradigmatic demonstrations of expert skills and activity may be enough for a system to gain competency in performing such tasks and in solving sequential decision problems. Such a system should explicate facets of the knowledge acquired and must have the ability to provide comprehensive explanation for how solutions are reached along trajectories of states reached, how objectives are achieved, and when/why it fails.
We need to investigate what the principles or paradigms of building an interpretable/explainable DRL system are: Is it enough to add some interpretability/explainability modules to an DRL system (as we would do in any ML system), or do we need to design a system from the early phases to be explainable (as it is conjectured to be the case for building collaborative and ethical systems)? In other words, do we need to design and develop inherently explainable DRL methods? An answer to this question may vary between domains and conditions of systems’ deployment, but this decision should not be based on any assumption about a trade-off between accuracy in prediction and interpretability. While authors in (PrinciplesPracticeofXML, ) propose methodological aspects for reaching such a decision for explainable machine learning, authors in (InterpretableML, ) conjecture that closed boxes are generally unnecessary, given that their accuracy is generally not better than a well designed interpretable model (whose probability of existence is rather high). However, a DRL method may comprise more than one interpretable models whose functionality is intertwined, and whose interpretability / accuracy may affect final outcomes.
We need to answer “what” information and “how” it should be communicated for providing effective explanations. We may take advantage of research in other scientific fields (social and cognitive sciences, as well as in philosophy) regarding the qualities of “good” explanations. Towards this, we need to consider not only the qualities of the explanation content (“what”), but also the modalities that will be used to provide surface representations of that information (“how”), the characteristics of the context in which explanations are provided, and the purpose that they need to serve w.r.t. constraints and requirements (pragmatic aspects). E.g., humans may be acquainted with specific modes and forms of information, using specific nomenclature and/or visual means, also in relation to responsiveness, temporal requirements in certain contexts, cognitive load, etc.
We need concise experimental protocols for explainability, considering the purposes and contexts of explanations’ provision together with human factors, in relation to the explanation problem and the required qualities of explanations. This will allow the community working on explainability to compare different approaches, and build a firm background, making conjectures and refuting theories on explainability, proceeding on more advanced results. While this holds for ML methods in general (ExplanationsQuality, ), experimental protocols are tremendously important for DRL methods, which have to support humans to infer faithful models of behaviour, objectives and capabilities of agents, towards developing trustworthiness in real-world problem solving settings.
Motivated by these challenging issues, in this article we review explainable DRL (XDRL) methods. We aim to understand their potential capabilities and limitations to solve interpretability/explainability problems while focusing on the needs of human operators. Operators need to be aware not only of the particular situation the agent faces, but also of the role that certain real-world features play in decision making, of the agent’s objectives, decision making capabilities and preferences to act, of the overall policy, and of the choices made under specific (potentially, critical) circumstances. In general, operators need to be able to understand and judge in an informed way when to trust an agent, when to take control and/or apply appropriate mitigation actions, and when to provide further advice for refining agent’s knowledge. While surveys on XRL exist (e.g. (SurveyXRL, ),(Multidisciplinary, )), here we aim at a comprehensive survey focusing on XDRL, providing details about all (according to our knowledge) works describing and providing well-justified insights on state of the art methods, under a common framework, distinguishing clearly between existing paradigms, providing details on evaluating these approaches, also with respect to transparency, and stating emerging and important challenges.
Structure of this article: In Section 2, we provide preliminary knowledge and discuss interpretability / explainability aspects that we must consider, as well as explainability desiderata. Section 2 formulates the problems of providing explanations for DRL agents. Based on the definitions provided, Section 3 identifies the functionality of the necessary components of an XDRL method and attempts a categorization of XDRL methods, based on the explanation paradigm they follow and needs they do address. Section 4 provides a comprehensive review of the state of the art methods, and Section 5 concludes the paper with final conclusions and challenges ahead. The Appendix (supplementary online material) of this article provides a detailed categorization of reviewed methods in three Tables, using a comprehensive set of characteristics.
2. Interpretability, Explainability and Transparency of DRL methods
This section sets the ground to proceed towards formulating the explainability problems for DRL agents, and providing a general framework of XDRL agents in the next section. First, it provides DRL preliminary knowledge, so as to introduce terminology and symbols that we will use in subsequent sections, assuming that the readers are acquainted with DRL methods. It proceeds to clarify the notions of explainability, interpretability and transparency, which is useful towards understanding the various facets of the problem, and states explainability desiderata. Finally, it provides definitions of the expainability problems for DRL agents.
2.1. DRL Preliminaries
Given a set of states and a set of actions available to an agent acting in an environment , possibly with continuous parameters, and a reward function , the goal of the agent is to learn a policy , i.e., a mapping from states to actions or to a distribution over actions, that will maximize its expected cumulative rewards.
Policy and reward models and are parameterized by variables’ vectors and , respectively, and take as input subsets of , a set of real-world features. A trajectory in time horizon (length) , is defined as a sequence of states and actions , for . The policy and reward parameters determine the optimal trajectory in the environment :
where is the discount factor between 0 and 1 that controls the trade-off between immediate and future rewards, and denotes the set of all possible trajectories in environment , given parameters and reward function . Those are the trajectories that can be sampled from the policy (i.e., ).
The goal of DRL methods is, given the reward function to tune the parameters so as to maximize the expected cumulative reward while acting in the environment , i.e., determine , such that
The agent, depending on the RL method used, may be required to fit a model of the environment and/or estimate the values of states, or the values of state-action pairs (i.e., the expected cummulative reward from that state, or state and action, respectively), or even (e.g in case of imitation learning) to learn a model of the reward function, tuning parameters , given the set of states’ features .
2.2. Explainability, interpretability and transparency
Given the above specifications and the models that a DRL method comprises (policy, reward models, etc - these are detailed in subsequent sections), we need to specify what it means to enhance a DRL method with explainability and interpretability, providing transparency in decision making.
Lipton (Lipton, ) asserts that explanation is post-hoc interpretability. Biran and Cotton (BiranCotton, ) (among others, e.g. C. Rudin at al. in (InterpretableML, )) define interpretability as the degree to which one can understand the response of a system. Relevance, of information for relationships contained in the data or learned by a machine learning model, for a particular audience into a chosen problem, is an important aspect of interpretable models in (Murdoch22071, ). Explanation is a particular way of obtaining such an understanding, selecting “what” must be communicated and “how” this is presented (e.g., using textual and/or visual means, providing “global” vs “local” or case-specific explanations, at different levels of detail, scale and granularity, etc.): This usually involves filtering interpretations and realizing surface representations of interpretations, using an explanation logic that exploits interpretable models. A distinction between “what” and “how” is also considered in (SurveyXRL, ),
To avoid overlaps and circularity and provide a concise definition of terms, close to the definitions from (SurveyXRL, ), we use the term “” to denote the surface representation of an interpretation provided to the user. The term “” is used to denote the explanation content provided by an interpretable model. Therefore, “” is the ability of a system to provide explanations’ content exploiting an interpretable model, “” is the ability to provide surface representations of interpretations. “” is the ability of the system to generate explanations that are understandable in the context of system deployment, w.r.t. domain-specific set of constraints: This definition of transparency is close to the first principle (definition) specified in (InterpretableML, ).
More specifically, considering that explainability is the ability of systems to provide qualitative understanding of responses and objectives, given environment features from the set and information provided by interpretable models, and close to our focus on DRL, we need to elaborate and clarify various aspects:
First, what does “qualitative understanding” mean; why do we require qualitative understanding; what should be made understandable, and how can this be achieved?
Then, what is a system’s response in the context of DRL methods? Do we consider any single such response at any time instance, or sets of responses (maybe aggregated or summarized) towards the evolution of a (sub-)trajectory, reaching certain objectives?
What features from the set of available features in should be considered for providing an explanation and how do we determine the importance of features?
Finally, a critical dimension that should be considered while elaborating on any of the above questions is whose needs an XDRL method must satisfy?
First, qualitative understanding expresses how humans understand the qualities of an agent’s policy, agent’s preferences, abilities and objectives, being able to provide accurately the rationale for agent’s chosen courses of actions, as the environment evolves within a state space, or for specific responses under certain circumstances. Such an understanding must occur in a human interpretable domain including domain specific concepts, entities and environment features that are presented using domain terms and other information modalities that humans are acquainted with while operating in a domain (e.g., sounds, or special visualization symbols). Such an understanding is in contrast to providing, for instance, detailed calculations, or detailed numerical assessments of any kind, e.g, by means of visualizations, at any scale and level of detail: This may of course be useful for understanding the inner workings of any method and model quantitatively, but it cannot serve the purpose of explaining the qualities of an agent’s decision making in an explicit way. This requirement for qualitative understanding is more relevant to domain operators who need to understand why an agent takes, or why it does not take a decision in a specific circumstance, at any level of detail, granularity or scale (we explain these terms further in the next paragraphs). Thus, such an understanding supports humans to generalize beyond individual explanations, allowing them to infer how probable it is for the agent to follow specific modes of behaviour, apply specific skills111Skills comprise options, macro-actions, behaviour modalities: Skills drive, in a hierarchical way, agent decisions at a low level of detail. or take a certain action under specific circumstances.
This is equivalent to supporting humans to infer models on (a) how the agent responds in certain circumstances w.r.t its abilities and preferences, (b) the agent’s behaviour w.r.t. its objectives, (c) the agent’s behaviour w.r.t its abilities, preferences and objectives, without considering “where do the numbers come from”. These models can be denoted respectively by , or , where denotes the agent’s response to the environment (e.g. an action, a (sub-)trajectory, or any skill), revealing agent’s behavioural choices at different scale and granularity levels, and provides a representation of those model elements that drive agent’s response at any time point or time period, at a low level. In other words, can be considered as a “window” to the DRL agent’s inner models’ aspects. Such functions may, for example, provide values, importance/saliency of state features assessed by the models, policy rollouts, even MDPs components or models’ raw weights. To keep the presentation simple, we consider a single function, whose indicated arguments or indicate which of the DRL models’ elements exploits.
At this point we can elaborate further on the different roles these models play:
models how humans interpret the agent’s activity under various circumstances in , based on their understanding of agent abilities and preferences. E.g., one may infer an agent’s preferable mode of behavior in , considering a subset of parameters .
models how humans interpret the agent’s activity in the environment, given their beliefs on the objectives targeted in . This allows humans to infer how the agent’s objectives affect its responses. E.g., one may infer the agent’s mode of behaviour given various weighted mixtures of features in , according to importance.
Finally, provides a model of how humans interpret agent’s activity based on their beliefs on agent abilities, preferences and objectives targeted in . E.g., one may infer an agent’s mode of behaviour given different combinations of capabilities, preferences and environment features’ importance.
Just to give a further example on these models and on their differences, given an apprenticeship learning paradigm, depends mainly on the training examples demonstrated to the agent; depends mainly on the set of features selected to train agent’s reward model, while depends on both.
The second question concerns the granularity and scale at which the agent’s responses are explained. While refers to the level of detail that the agent behaviour is considered, refers to the extent where a specific response is considered. Thus granularity may refer to the actions performed at any time instance, or -in a more coarse level- to the trajectories chosen and their aggregated features, or - in even more coarse terms- to the mode of behaviour or skill exercised by the agent within a time period. Scale refers to the temporal or to the spatial (in whatever space dimensions) extent in which responses are considered. For instance, one may consider a trajectory within a specific time horizon as a response. Sub-trajectories in more restricted time horizons may be considered, while one may consider trajectories in different spatial extents and dimensions. Thus, depending on the kind of response, may be instantiated to either: (a) , denoting specific actions, (b) , denoting specific (sub-)trajectories, or (c) , denoting specific skills, given a set of skills , where belongs. For instance, models how humans understand an agent’s choices of skills exercised in , based on their understanding of agent’s preferences and objectives, using the elements provided by . Finally, it must be noted that granularity and scale parameters are not included in the denotation, so as to simplify it. For instance, an action is a response given at a specific time instant and it is denoted by , while a trajectory in a time horizon , from (now) to , is denoted by , etc.
The third question refers to the environment features in that are being used for providing an explanation of agent’s responses. This is important, as various subsets of features in may be chosen, or the features in can be used in different ways to provide an understanding of agent responses: Scale and granularity directly affect how the agent’s responses are perceived: For instance, considering scale, a trajectory may be considered in a space of or less dimensions, given different projections of the trajectory in subspaces. As a concrete example, consider a trajectory in the 3D space, where : One may view such a trajectory in any 2D projection, thus considering responses in any subset of features in . Generally, we may consider the “projection” of responses in any subset of . Granularity may impose filtering and/or aggregation of features in the temporal dimension. On the other hand, a critical aspect to be considered about features in is how the agent’s responses are affected, if any state of is projected in a ? This is in contrast to the simple response projection mentioned above, as it affects how the agent perceives the environment: Such “masking” of features and “perturbations” reveal causalities and sensitivities on environment features.
The last question concerns whose needs does explainability addresses. As already pointed out, in this article, we focus on the needs of human operators, who require to solve domain problems in real-life contexts, rather than on the needs of machine learning and RL experts. This is connected to several aspects considered above, such as (a) how humans’ beliefs and understanding of agents’ inner workings are modeled, (b) how and in which terms humans are assumed to perceive and understand the DRL inner workings, (c) in which scale and granularity DRL responses are provided, (d) how sensitivity on features and features’ importance is communicated in specific circumstances and overall, and (e) how, specifically, the policy and/or the objective function are made understandable (also in connection to (b)).
In this article, we do not elaborate on issues related to human factors and modelling of humans, nor on the effectiveness of different surface representations of explanations. There are many works exploring these issues and it is out of this article’s scope.
2.3. Explainability desiderata
An agent, to be explainable, needs an interpretable model and an explanation logic. The interpretable model provides explanations’ content, while the explanation logic selects the modes and the way explanations are surfaced, towards meeting transparency requirements. Interpretable models may be functional parts of the agent, i.e., the agent may have the ability to provide responses without them, as we will explain in subsequent sections.
At high levels of automation, the transparent agent must be trustworthy, in the sense that operators should be comfortable relinquishing control to it, as pointed out by Z. C. Lipton in (Lipton, ). A major factor to that is the accuracy of predictions and “goodness” of prescribed actions in terms of solving the identified problems (predictive accuracy), w.r.t the interests of stakeholders (e.g., without increasing cost of operations or compromising safety). However this is not sufficient: Transparency is crucial, especially when system’s responses do not comply with experts’ usual practices or intuition, and in cases where operators need to understand system’s capabilities and limitations.
There are different desiderata for transparency and explainability, regarding human interaction/communication issues, and aspects concerning properties of the interpretable models. This becomes apparent if we consider that transparency requires explainability that meets pragmatic constraints, which in its turn requires interpretable models that provide information towards enhancing humans’ understanding of systems’ responses, objectives and capabilities.
Thus, desiderata concern providing explanations that are (a) accurate on the information provided, (b) fitting the purpose and context of system deployment, (c) with respect to operational requirements (e.g., on the type of information provided per type of problem instance), operational constraints (e.g., time constraints and other constraints regarding privacy and security, etc.), and with respect to ethical aspects.
Towards satisfying these desiderata, the type and amount of information being provided and the way explanations are being surfaced are crucial. Scale and granularity of responses are important aspects. Specifically, concerning the type and amount of information provided, a system may provide a systematic trace of its inner-workings (low scale), with an exhaustive enumeration of causal paths and features being considered per problem instance (low granularity), increasing the complexity of explanations, which may also be affected by the size and complexity of the interpretable model. The ways in which this information is communicated affects transparency and trustworthiness: People need to accomplish tasks with awareness of the problem(s) faced, also with respect to pragmatic constraints and ethical issues (e.g., to sensitivity of information, to potential dual-use of information and system, fairness - making no discrimination against groups of entities) and with respect to any contextual factor that affects acceptability of system’s responses and explanations. For instance, a lengthy explanation regarding a system response to an emergent problem instance (with details known to the explainee), may lower explanations’ acceptability and can determine system use in contexts where timely response is necessary, unacceptable. Selecting the most important causes/features that should be included in an explanation, focusing especially on those that are of interest to operators w.r.t the problem instance considered, is very important, as evidenced by studies regarding human cognition and explanation effectiveness (deGraafMalle, ), (Miller, ). This supports effective update of operators’ models on agent’s policy, objectives, preferences and capabilities, relating operators’ existing knowledge and agents’ interpretations.
Causality and robustness are additional features for the system to be trustworthy and these properties must be conveyed to humans via the explanations provided: While causality requires identifying the input features and causal chains that affect system’s responses, robustness requires that the system performance is not affected considerably with small variations of input features, especially of those that are not important to problem instances: Importance of features per problem instance, and sensitivity on features, are important elements for explanations.
Regarding accuracy, we distinguish predictive accuracy, and the accuracy of the information provided by the explanation logic (descriptive accuracy). While accuracy of predictions concerns the ML method used, accuracy of the information provided in explanations, concerns the capacity of the interpretable model and of the explanation logic to locate and present the exact models’ elements affecting agent’s responses. In cases where the interpretable model is distinct from the models used by the decision-making agent, then the interpretable model must replicate agent’s behaviour for explainability purposes with fidelity: The interpretable model must reproduce agent’s responses accurately, identifying also the important features that affect agent’s responses. While this may happen with local fidelity (i.e., in the vicinity of the problem instance being solved), the interpretable model may provide global perspective of the agent’s behaviour (e.g., explaining objectives, intended outcomes, preferences, capabilities). These important distinctions are clarified by the problem definitions stated in the next section.
2.4. Problem formulation
Based on the goal of DRL agents, as specified in section 2.1, and on the explainability / transparency goals and desiderata, as specified in sections 2.2 and 2.3, respectively, this section provides definitions for the model explanation, outcome explanation and model inspection problems for DRL methods, refining those provided in (Survey, ), so as to connect explainability for DRL methods to explainability of ML prediction methods, allowing a more comprehensive view of explainability in ML: Overall, the model explanation problem aims at providing explanations regarding the overall, global logic behind agents’ responses, revealing also global information on agents’ preferences and capabilities. The outcome explanation problem is related to explaining responses provided by agents under certain circumstances, specifying the correlation between responses and values of features in , at a low level of granularity (i.e., regarding certain actions) and scale (i.e., in restricted temporal and spatial extent). Finally, the model inspection problem concerns the provision of information that reveals aspects and properties of the individual models that a DRL method comprises.
Definition 2.4.1 (Model Explanation Problem). Given a DRL method, the model explanation problem consists in finding a global explanation of behaviour, belonging to a human interpretable domain through an interpretable model , in a specific environment , using that provides a representation of policy and objectives models’ elements. The global explanation is provided via an explanation logic that reasons over .
Providing a solution to the model explanation problem aims at updating , interpreting agent’s responses at coarse levels of granularity and at large scale extents, exploiting agent preferences, abilities and objectives in . We may say that solutions to the model explanation problem aim to provide global (i.e., independently from specific states) explanations both for the policy and for the objectives of the agent.
Refining Definition 2.4.1 to either the policy or to the objectives models of a DRL method, we have the policy model explanation problem and the objectives model explanation problem:
Definition 2.4.2 (Policy / Objectives Model Explanation Problem). Given a DRL model, the policy (objectives) model explanation problem consists in finding an explanation of behaviour, belonging to a human interpretable domain through an interpretable model for policies ( for objectives), using a representation of agent’s policy or objectives models’ elements, provided by . The global policy/objectives explanation is provided via an explanation logic (respectively, ) that reasons over (respectively, ).
It must be noted that in the above definition, the policies’ interpretable model includes optionally (this is what square brackets denote) a representation of agent’s objectives: We leave this as an option, given that the objectives may be assumed to be known, or interpreted separately.
Definition 2.4.3 (Outcome Explanation Problem). Given a DRL method, the response explanation problem consists in finding a local explanation , belonging to a human interpretable domain through an interpretable model at state , using a representation of agent’s policy and objectives models’ elements, provided by . The local explanation is provided via an explanation logic that reasons over and .
It must be noted that the outcome explanation problem concerns skills and actions that the agent applies at states , as well as (sub-)trajectories that at time are in state . In any case, the response explanation problem concerns agent’s responses in fine granularity and small scale. Additionally, it may also concern providing local interpretation of the objective function, motivating some responses instead of others.
Definition 2.4.4 (Local Response/Objectives Explanation Problem). Given a DRL method, the local response/objectives explanation problem consists in finding an explanation , belonging to a human interpretable domain through an interpretable model for responses (respectively, for objectives) at state , using representations of models’ elements provided by . The local explanation is provided via an explanation logic (respectively, ) that reasons over (respectively, ).
Providing a solution to the outcome explanation problem aims at updating a model , towards interpreting agent’s responses and objectives at fine levels of granularity and scale, at any state. We may say that any solution to the outcome explanation problem aims at providing local joint explanations both for the fine grained responses and objectives of the agent, at particular states.
The model inspection problem for DRL methods concerns providing surface representations of DRL internal models’ elements that would reveal agent’s internal models’ workings and properties, provided via the function .
Definition 2.4.4 (Model Inspection Problem). Given a DRL method, the model inspection problem consists in finding a surface representation of DRL models’ elements, or of models’ properties, exploiting representations of models’ elements provided by . The notation indicates that exploits either , or both, in a joint manner.
In contrast to providing model explanations, the model inspection problem concentrates on the presentation of individual models’ elements and properties at low levels of detail, towards inspecting the inner workings of models, without necessarily revealing an understanding of agent’s policy, objectives, or responses.
3. Framework and paradigms for XDRL methods
Given the above stated definitions of the DRL explainability problems, Figure 1 provides the overall architecture of a framework for XDRL methods: In addition to the inner dark box, which provides the main components of any DRL method222This comprehensive DRL blueprint has been suggested by S.Levine, according to our knowledge., Figure 1 presents the basic pipeline for XDRL. DRL models provide interpretations, denoted by , which, moving from dark to lighter areas of the blueprint, are exploited to provide content to an explanation logic. It must be noted that, as defined above, any such model exploits either , or both, in a joint manner.
The policy of a DRL agent is represented by the policy model with parameters (e.g., a deep neural network (NN)). Given environment states’ features in , this model provides a probability distribution on all actions that can be applied at a state. In conjunction to that model, there can be other models either fitting a value function (e.g., on states or on state-action pairs), or modelling the way the environment evolves, or fitting the reward function. Figures 1 and 2 show only the reward model, placing special emphasis on the objectives of the agent. This model approximates the reward function at any time instance , given a state-action pair . Overall, the DRL method exploits the reward model (possibly, in conjunction to other models) to fine-tune the policy model parameters , while the agent may generate trajectories’ samples in to test the optimality of the policy w.r.t. the objectives.
![]() |
![]() |
Figure 1. Overall framework for XDRL | Figure 2. Framework for XDRL |
with interepretable models |
To inspect individual models, function provides a representation of policy model and of objectives model elements. This may solve the model inspection problem, but it does not solve any of the (global policy or objectives) model or (local responses / objective) outcome explanation problems.
Figure 2 shows a refined version of the XDRL framework with interpretable models exploiting the representations provided by . As it is shown, any such interpretable model may be an objectives model (), a local responses model (), or a policy model (). Any XDRL agent may contain all or any combination of these interpretable models, at the local and/or at the global level.
Explanations are generated by functions that exploit the interpretable models, maybe in combination with the elements provided by the model inspection components. Figure 2 shows functions providing explanations for objectives (), responses () or for the policy model ().
3.1. Interpretable methods
![]() |
![]() |
Figure 3. Overall framework for the interpretable box design paradigm | Figure 4. The interpretable box design paradigm |
with specific interpretable models |
Some of the models that a DRL agent exploits may be directly interpretable: For instance, the policy model may be a decision tree, or the reward model may be a linear function on features in . In this latter case, realizes the weights of the reward function , where are functions on features in and action parameters. As functions become arbitrarily complex, they hinder the interpretability of the reward model.
Interesting variations of the XDRL framework provided in Figure 1, are frequently met in the literature, where interpretable models either (a) substitute inner constituent models of the DRL method (thus making the DRL method “inherently” interpretable), or (b) are being trained using the DRL agent and its models as oracles in order to provide explanation content. These cases, in close relation to definitions provided in (Survey, ), aim to solve the transparent box design problem. In this paper, these variations are considered to offer distinct paradigms for building interpretable models, according to what we call the interpretable box design paradigm333We do not use the term “transparent”, as we associate transparency with pragmatic aspects regarding the provision of explanations..
Interpretable models, as Figure 3 shows, are parameterized using variables’ vectors for policies and for rewards (possibly in conjunction to other models), distinct from the parameters of the DRL models, if separate DRL models exist. It must be noted that interpretable models in these cases do not necessarily exploit DRL models’ elements by means of . As a consequence, while model inspection facilities may be provided for the constituent DRL models, these in some cases are totally disconnected from the interpretable models.
Interpretable models using the DRL method as an oracle (the second variation mentioned above), can be trained either during or after training the DRL models. While training interpretable models during training the DRL model may result to instability and inefficiency of the training process, aiming to reach a moving target, it may provide insights to the DRL learning process. The training process may use samples and results (e.g., rewards, state/action values) provided by the deep method, while it may also exploit the models of the DRL method in a direct manner, delving into their inner workings. This results into two distinct paradigms that refine the interpretable box paradigm: The mimicking and the distillation paradigms, respectively.
It must be noted that although the distillation process has been introduced as a NN model compression process through training a student NN, and has been introduced as a term in (HintonDistillation, ), the first work that introduces it for sequential decision making (as a model compression technique, not as a method for building interpretable models) is described in reference (PolicyDistillation, ), where the teacher provides samples, each comprising an observation sequence and a vector of state-action -values.
For DRL methods, the distinguishing line between the distillation and the mimicking processes is not that clear in the literature: Here we consider these as two distinct paradigms for providing interpretability to DRL methods.
3.2. Distillation and mimicking: Two distinct paradigms for interpretability
![]() |
![]() |
Figure 5. The overall framework for distillation | Figure 6. The overall framework for mimicking |
The distinct paradigms are shown in Figures 5 and 6. The main difference between the two paradigms concerns the “receptive field” of the interpretation process: As Figure 5 shows, the distillation process gathers knowledge from the trained DRL agent, exploiting any of the constituent DRL models in a direct way through a process of transforming DRL models’ elements into interpretable models. In contrast, the mimicking process, as Figure 6 shows, monitors the interaction of the RL agent with the environment, and gathers interaction samples, recording agent decisions, state transitions, rewards and consulting DRL assessed values.
As an important note to the above, one may argue that the architecture shown in Figure 5 is more general than the one shown in Figure 1. Actually, the architecture shown in Figure 1 abstracts all XDRL paradigms while providing exact interpretations of DRL models. The two paradigms offer interpretable models that are approximations of the DRL constituent models: The distillation paradigm shown in Figure 5 aims at exploiting the knowledge acquired by the DRL models via model inspection facilities, while the mimicking method in Figure 6 produces interpretable models without exploiting the inner working components of DRL agents.
4. XDRL methods: State of the Art
This section aims at a detailed presentation of XDRL approaches. For each of the reviewed approaches, we describe the overall approach and motivating points, assumptions and important technical details, and we conclude with important aspects w.r.t. the specific problem it addresses, paradigm it follows, interpretable models it constructs, required access to constituent closed-box models, and finally, information on evaluation objectives, methods and results. In addition to these, the Appendix summarizes the XDRL approaches described, focusing on important aspects of their functionality and applicability, using a comprehensive list of characteristics that includes those described above, and additionally, the XDRL requirements for the DRL methods used, as well as “XDRL - DRL interplay aspects”: Access to DRL models’ information and/or samples required, in an online mode or in an offline mode.
As a first note before delving into XDRL methods, there are many proposals for interpreting deep NNs models, mainly through distillation and mimicking approaches. The main emphasis of these approaches is building high-fidelity interpretable models, in comparison to the original models. These approaches differ in several dimensions: (a) the targeted representation (e.g., decision trees in DecText (ExtractingDTfromTrainedNNs, ), logistic model trees (LMTs) in reference (LMTExtractionfromANNs, ), or Gradient Boosting Trees in reference (interpretableDeepModelsforICU, )), (b) to the different splitting rules used towards learning a comprehensive representation, (c) to the actual method used for building the interpretable model (e.g., (LMTExtractionfromANNs, ) uses the LogiBoost method, reference (ExtractingDTfromTrainedNNs, ) proposes the DecText method, while the approach proposed in reference (interpretableDeepModelsforICU, ) proposes a pipeline with an external classifier, (d) on the way of generating samples to expand the training dataset. These methods can be used towards interpreting constituent individual DRL models employing (deep) NNs. We do not aim to review these in this article, given the thorough review provided elsewhere (e.g. in (Survey, ), (Murdoch22071, ), (PrinciplesPracticeofXML, ), (InterpretableML, )). However we delve into XDRL approaches that exploit any of these methods.
4.1. Solving the model inspection problem
Here we provide a succinct review of prominent general methods addressing the DRL model inspection problem.
4.1.1. Why should I trust you?
M. T. Ribeiro, S. Singh and C. Guestrin in (LIME, ) propose LIME for explaining the predictions of any classifier in an interpretable and faithful manner. LIME learns an interpretable model locally around the prediction, searching for instances in the vicinity of the original instance through perturbations of uniformly sampled features in subsets of (i.e., non-zero elements). Emphasis is given to the qualitative understanding of correlations between input features and model responses, aiming to guarantee local fidelity, explaining how the model behaves in the vicinity of the instance being predicted, using a proximity measure. Thus, while LIME addresses the model inspection problem assuming binary vectors for instances’ interpretable representations, it contributes to the outcome explanation problem solution, as it provides the importance of features in a local only context.
Building on explanations of individual predictions, authors in (LIME, ) propose the SP-LIME method by presenting representative individual predictions and their explanations in a non-redundant way, framing the task as a sub-modular optimization problem. SP-LIME provides a global perspective and contributes to solving the model explanation problem.
Experiments performed with simulated users regard (a) the faithfulness of explanations to the model, given a set of gold features that explanations must recover; (b) the trust on predictions, given untrustworthy features in predictions; (c) the trust on the model, given a set of trustworthy examples covering cases with varying subsets of important features. Evaluations with human subjects regard the accuracy of choosing the best classifier, and improving the classifier by eliminating features which are untrustworthy (measuring classification accuracy after rounds of interaction).
LIME provides a generic (model-agnostic) method that contributes to local and global (model) interpretability. In the context of DRL, it can be used for explaining any of the models (e.g., to explain local decisions for selecting an action given an interpretable representation of states’ features), although not being designed for that particular task. Functions provide binary vectors for instances’ representations, while (for outcome explanation ) can theoretically be any interpretable model, e.g., a linear model, a decision tree, etc. Surface representations of instances’ interpretable representations (for model inspection, via ), as well as explanation logic for local explainability and for model explainability, are provided using visual means.
4.1.2. Shapley Additive Explanations
The Shapley Additive Explanations (SHAP) approach aims to provide a unified approach to interpreting model predictions, showing how interpretation methods are related, and when one method is preferable over another. S. M. Lundberg and L. Su-In in (SHAP, ) identify the class of additive feature attribution methods with a linear function of binary variables as an explanation model: , where , the number of simplified input features and .
There is a family of methods - including LIME, DeepLIFT (DEEPLIFT, ), Layer-Wise Relevance Propagation (LayerWiseRelPropagation, ) and classic Shapley value estimation methods- that are additive feature attribution, and authors in (SHAP, ) show to unify six of the existing methods. They provide theoretical results on the existence of a unique solution in this class of methods with a set of desirable properties, that several recent methods in the class lack: Local accuracy, requiring the explanation model to at least match the output of the original model; missingness, representing features presence so as to locate the features with no prediction impact on a local decision; and consistency, regarding the relation of input’s contribution and input’s attribution.
Specifically, considering a simplified input mapping to the original input through the mapping function : , and input , with to be the set of non-zero indexes in ( has missing values for features not in ) s.t. , SHAP values are the Shapley values of a conditional expectation function of the original model , given the model prediction . SHAP focuses on explaining a prediction , given a model with input and a simplified input . Such local methods ensure that , whenever , where is the local interpretable model (for outcome explanation). SHAP, similarly to LIME, provides an interpretable model as a linear function of binary vectors for instances’ representations provided by . Surface representations of instances’ interpretable representations for model inspection, as well as explanation logic for local explainability, are provided using visual means.
Based on insights from unifying methods, authors in (SHAP, ) present methods that show improved computational performance and/or better consistency with human intuition than previous approaches. Evaluation of Kernel SHAP vs. LIME and Shapley sampling values show improvement in terms of learning accurate estimates with fewer evaluations of the original model (sample efficiency). User studies in a limited setting show that SHAP values prove more consistent with human intuition than alternative feature importance allocations represented by DeepLIFT and LIME, which fail to meet the desirable properties identified. Finally, authors use MNIST digit image classification using a convolutional network to compare SHAP with DeepLIFT and LIME, showing that SHAP is able to explain differences in classifying a digit to a class rather than to a near-class, measuring the probability of classifying instances to classes and changes in log-odds when masking over images.
A concrete example of employing SHAP (actually, a specific instance of LIME, named Kernel SHAP) in the context of DRL, is provided in reference (XRL4TrafficSignals, ).
This article uses Policy Gradient to implement a traffic signal control agent on a signalized
roundabout. Kernel SHAP estimates the effect of the road
detectors state on the agent selected phases, interpreting the decision taken, in relation to the traffic volumes and the
lanes occupancy.
Other methods, such as for instance Layer-wise Relevance Propagation (LRP) (LayerWiseRelPropagation, ) and Neighborhood Component Analysis (NCA) (NCA, ) can be used for addressing the model inspection problem for any of the constituent DRL models, also providing (a) importance of features in , and (b) reduction of predictors’ dimensionality, which can facilitate data visualization and fast classification. Such methods are thoroughly reviewed in (Survey, ).
Following this short presentation of model inspection general methods, we proceed to review XDRL methods addressing any of the model, outcome explanation, or the interpretable box design problems.
4.2. Solving the policy explanation problem
4.2.1. Summarizing agent behaviour to people with HIGHLIGHTS
With the goal of increasing people’s familiarity with agent’s capabilities and limitations, D. Amir and O. Amir in (Highlights, ) propose the HIGHLIGHTS and the HIGHLIGHTS-DIV algorithms that choose trajectories highlighting important and distinct aspects of agent behaviour. In doing so, this approach provides an overview of an agent’s policy, rather than explaining specific responses.
Specifically, the HIGHLIGHTS online algorithm constructs summaries of agent’s behaviour by selecting (sub-) trajectories with important states. A state is considered important if taking a “wrong” action in that state (i.e., an action that is not prescribed by the policy) can lead to significant decrease in future rewards. To provide context to the user, for each important state the algorithms extracts a sub-trajectory including neighboring states (before and after the important states) and actions. In order to highlight behaviour in various sub-spaces in the state space, HIGHLIGHTS-DIV avoids including in the summary (sub-) trajectories with very similar states, if these are equally important.
HIGHLIGHTS has been evaluated by playing the Pacman game, based on the observed behaviour rather than scores. Q-values in this setting are defined as a weighted function of state feature values. Given different agents trained for different number of episodes, human participants were asked to select the best agent based on the summaries provided, while subjective opinions about the helpfulness of summaries were elicited, given summaries of the same agent. The surface representation of the explanation in (Highlights, ) is provided by means of video clips, but in general, other means and information modalities may be used (e.g., as in (ContrastiveExplanations, ) that is presented in a subsequent section). This evaluation resulted in observations regarding participants confidence in choosing agents based on HIGHLIGHTS summaries. Experiments show that short summaries may not enable participants to correlate their confidence with their ability to assess agent’s abilities. However, with statistical significance and for very well trained agents, participants preferred summaries generated by HIGHLIGHTS over baselines, and summaries generated by HIGHLIGHTS-DIV over HIGHLIGHTS.
Overall, this approach addresses5 the policy model explanation problem, demonstrating agent behaviour. It is important to point out that this approach by-passes the construction of an interpretable model for policies, but it gathers explanations’ content by selecting examples of agent’s behaviour in important states. The representation of importance is realized via the function which has access to agent’s policy and Q-values.
T. Huber et al. in (LocalAndGlobal, ) use a version of HIGHLIGHT-DIV in a combination of global and local explanations for DRL agents. Specifically, they use a version of the Layer-wise Relevance Propagation (LRP) (LayerWiseRelPropagation, ) method, i.e. LRP with the argmax rule (LRP-argmax, ), to augment trajectory summaries with saliency maps. This local approach is presented in detail in section 4.4.11. LRP has been used to show what information is important to the agent towards a specific decision: The approach presented in (LocalAndGlobal, ) addresses the policy and the responses explanation problems by combining methods, contributing to our understanding of the role of saliency maps in the context of explaining agents’ behaviour: While saliency maps have been shown to improve classification decisions in images (albeit with 60% of correctness in (Alqaraawi4, )), their role in RL is not that clear (e.g. as shown in (TransparencyandExplanationInDRL, )). However, they may provide significant positive effects when combined with other methods (as in (Erwig23, )). Empirical evaluation of the proposed combined method aims to investigate (a) the mental model of humans about the DRL agent (via a retrospection task), (b) humans’ ability to assess agents performance (via an agent comparison task) and (c) participants satisfaction with respect to the explanations presented (via answering explanation satisfaction questions). The results of this study reinforce prior findings for HIGHLIGHTS-DIV. Overall, in this study, the choice of states shown to humans was more important than the inclusion of local explanations in the form of saliency maps. For saliency maps, the results reported are mixed: There were no significant differences between the saliency and nonsaliency conditions in the study. Showing saliency maps as part of a video is more challenging for humans, in comparison to showing them in a still image. As authors point out “[saliency maps] in RL [add a] layer of complexity as interpretation also requires making inferences regarding how the highlighted regions affect the agent’s long-term, sequential decision-making policy”. However, in combination with behaviour explanations, saliency maps put the important features in context, while they are not preferable when humans get highlights of agents’ behaviour.
Extending the HIGHLIGHTS algorithms, P. Sequeira and M. Gervasio in (InterestingnessElements, ) propose using not only state importance, but various interestingness elements so as to highlight important agent’s behavioural aspects, increasing humans’ understanding of agent’s capabilities and limitations. Such elements are (in)frequent situations, (un)certain executions, observation minima and maxima, and most likely sequences to maxima, each serving a specific explanation provision purpose. To compute these elements, the proposed framework collects data from agent’s interaction with the environment, such as the number of times the agent has specific observations, estimated transition probabilities, as well as estimated Q-values and V-values. Although there are elements (e.g., frequency) that may provide a good understanding of agent’s behaviour characteristics, overall, results show that no single summarization technique — and hence no single interestingness element — provides a complete understanding of every agent in all possible situations of a task. Ultimately, different combinations may be required for agents with distinct capabilities and levels of performance.
4.2.2. Contrastive explanations for reinforcement learning in terms of expected consequences.
Motivated by findings indicating that contrastive explanations offer intuitive motivations for humans to understand why one performs an action instead of another (Miller, ), J. van der Waa et al. in (ContrastiveExplanations, ) propose a method for providing contrasting descriptions of rollouts generated by an RL agent’s policy and a policy that is driven by humans’ (non-experts in RL) queries.
This proposal (a) defines a translation of states and actions in a set of descriptive states’ classes and action outcomes, by training binary classifiers, similarly to (AutonomousPolicyExplanation, ), and (b) proposes a method for obtaining a foil policy based on the foil in the user’s query, suggesting potential alternative actions in a particular state and consecutive state-action pairs. This foil policy, based on an update of the function to reflect user’s preferences (and an informed update of the reward function) is contrasted to the fact policy learned by the agent. The explanations are based on generating rollouts of specific length simulating the effects of and of , based on the environment’s transition function and starting on a state before the state specified in the user’s query. State-action pairs in rollouts are transformed in descriptions which are then used for generating the trajectory contrastive descriptions by instantiating natural language templates.
Authors in (AutonomousPolicyExplanation, ) report on a case study using a simple RL setting with a very restricted objective, addressing either the next action of the agent policy or the entire policy: They show that human participants prefer explanations using the policy, as in most cases the next action is evident to the user.
This approach addresses the policy model explanation problem, by contrasting rollouts of the agent’s policy to foil policy rollouts. The interpretable model provides qualitative descriptions of rollouts by means of distinct state and action classes. The interpretation process accesses the MDP, which may be fitted by the learning process, the state-action Q-values learned, and the rewards received, via the function . Finally, the explanation logic uses natural language templates that are instantiated by qualitative (potentially, contrasting) descriptions of rollouts.
4.2.3. Establishing appropriate trust via critical states.
Motivated towards helping end-users to build a mental model of agent’s policies and capabilities, S. H. Huang et al. in (CriticalStates, ) propose approaches for computing critical states in a policy-agnostic way, requiring only access to action-value function . This proposal is driven by the insight that for many tasks the essence of the policy is captured by agent’s reactions in a few critical states.
A critical state is one in which it is very important to take a certain action: As defined in (CriticalStates, ), varies greatly across different actions , i.e., is very high for some actions, but for most of them, this value is mediocre or low. Given agent’s policy, critical states are exposed to end-users who can spot false-positive or false-negative critical states (i.e., states that are considered as critical by the agent, but not by the humans and vice versa). Inspecting the critical states, the end-users may refine their trust to the agent, or gain proper knowledge on the situations where they have to take control. For instance, both false-negative and false-positive critical states indicate that the robot has failed to learn something fundamental about the task. Similarly, if the policy identifies a true-positive critical state but is mistaken about which action is correct in that state, then the end-user may not trust the policy.
This approach does not provide an explicit interpretable model but aims to provide a coherent view of agent’s capabilities and limitations so as to establish trust to the agent. Critical states are determined by accessing -values via the function . The interaction with human experts concerns (a) spotting false-positive, false-negative states, or incorrect-actions in critical states and assessing whether to deploy the agent, (b) deploying the agent which operates with the user in the loop, who she is able to take control whenever needed. A critical issue is whether observing a set of critical states leads participants to develop appropriate trust in high-entropy policies that generated those critical states, using the Soft Actor Critic algorithm (SAC). Also, while we may consider that this approach addresses the policy model explanation problem, it may be hard for the human to generalize and consider how the agent reacts in states that have not been presented as critical. Also, while the surface representation of states-action pairs is shown to be depicted using visual means, it is harder to do so in settings where the dimensionality of the state/action space is large.
4.2.4. Graying the Black Box: Understanding DQNs
The main observation driving the work of T. Zahavy, N.B. Zrihem and S. Mannor in (GrayingBlackBox, ) is that Deep Q-networks (DQN) is learning an internal hierarchical model of the domain without explicitly being trained to. Therefore, authors propose a methodology and tools to analyze DQN, aiming to identify spatio-temporal abstractions directly from the learned representation in two different ways: First, manually clustering the state space using hand crafted features, and second, computing the Semi-Aggregated Markov Decision Process (SAMDP), an approximation of the true MDP learned from data, by means of spatio-temporal abstractions. Visualization tools, including t-SNE that exploits directly the collected neural activations of states (a model inspection task), saliency maps, as well as analysis tools for understanding the common attributes of states’ clusters, and for analyzing dynamics between clusters, aim to provide facilities towards understanding, debugging and interpreting the policy model.
Specifically, authors in (GrayingBlackBox, ) explain the five stages towards building an SAMDP model aggregating states and skills, thus allowing analysis with spatio-temporal abstractions: (0) Feature selection; (1) State aggregation, mapping the MDP feature space to the abstracted MDP state space enforcing temporal coherency among trajectories; (2) Identification of skills (a.k.a sub-trajectories) from the data; (3) Inference of skill length, the SAMDP reward and the SAMDP transition probabilities; and finally, (4) selecting the best among the SAMDP candidates, according to evaluation criteria proposed.
Overall, this approach proposes a methodology and abstraction methods addressing the policy model explanation problem for DRL methods, where the policy is learned by the DQN method. The interpretable model is an SAMDP that is devised by consulting manually crafted state features, as well as rewards and neural activations of states provided by . The explanation logic is implemented by t-SNE maps with representative states per cluster, and appropriate t-SNE enabled SAMDP visualizations. In conjunction to that, authors demonstrate how to do (policy) model inspection by processing the network’s neural activity using visualizations enabled by t-SNE.
4.2.5. Towards Better Interpretability in Deep Q-Networks
R. M. Annasamy and K. Sycara in (betterInterpretability, ) propose an interpretable NN architecture for -learning, which provides explanations for visual agents. The aim is to provide an understanding of the policy model, visualizing clusters of state representations in attention maps, and using saliency maps.
The proposed model learns a latent representation of states that captures important visual aspects of input images, and an association between representations of states and representations of keys in a key-value store using an attention mechanism. Each key represents an action and an associated -value, among such values: Thus, the network learns to associate states to “representatives” of (action,-value) pairs, which serve as cluster ”exemplars”. The attention weights over keys and their corresponding value pairs are used to calculate -values. Different types of losses are being linearly combined to train the network: Bellman error, distributive Bellman error, reconstruction error and diversity (over keys) error. Uncertainty of the attention weights is being used to drive exploration during training, approximating an upper confidence on the -values. Through key inversion (which not a process driven by a certain input) authors attempt to find important aspects of the input space that influence the choice of a particular action-return pair.
As concluded in (betterInterpretability, ), with a directed exploration strategy, the proposed model can reach training rewards comparable to the state-of-the-art deep -learning models. Specifically, the best agreement between the actions selected using the distributions in the image space and latent space is rather low, but the explanations provide useful insight into the kinds of features extracted by convolutional layers: This possibly suggests, according to the authors, that reconstructions rely heavily on generalizability to unseen keys. However, results suggest that the features extracted by the NN are shallow, and the agent does not model interactions between objects.
Concluding, this approach addresses the interpretable box design problem for deep -learning methods, focusing on providing global policy explanations for visual agents. The interpretable model is an attention model associating state embeddings to key-value embeddings. The explanation logic provides keys - state embeddings clusters’ visualizations using t-SNE for any specific -value, as well as images reconstructed via key-value (i.e., action, -value) pair inversion.
4.2.6. Interpretable Off-Policy Evaluation by Highlighting Influential Transitions
Off policy evaluation (OPE) estimates the value of a policy using data collected under another policy. Aiming to enable human experts to assess and analyze the validity of off-policy evaluation estimates, O. Gottesman et al., in (InterpretableOPE, ) propose a hybrid human-AI system that highlights observations in the data whose removal will have a large effect on the OPE estimate. Their proposal aims (a) to address the needs of real-world applications where the data itself might never be enough; and (b) to involve domain experts in the integration of decision support tools, incorporating their expertise into the evaluation process. In so doing, they formulate a set of rules for choosing which observations to present to domain experts for validation.
Specifically, in this article authors propose a framework for using influence functions to interpret OPE, and discuss the types of questions which can be shared with domain experts to use their expertise in debugging OPE: Influence functions show the impact that the removal of a transition has on state-action values or on the value of the evaluation policy. Based on this framework, they develop computationally efficient algorithms to compute the exact influence functions for several importance sampling estimators, as well as two broad function classes for Fitted Q-Evaluation (FQE): kernel-based functions and linear functions.
The benefits of influence analysis are demonstrated on a cancer simulator, with the objective to identify limitations in the evaluation process and make evaluation more robust. Furthermore, authors present results of analysis together with practicing clinicians of OPE for management of acute hypotension from a real intensive care unit (ICU) dataset.
Overall, this approach proposes a framework that addresses off-policy evaluation independently from any reinforcement learning method. It is closer to the policy explanation problem, since the influence analysis proposed aims to interpret a policy evaluation method. The method can be applied to parametric and non-parametric fitted Q-evaluation models, regarding continue or discrete states and actions, although the presentation and evaluation considers discrete cases. The method focuses on the influence of specific state transitions to the value of the policy, and thus, it can be considered to be a method that reveals the contributions of local decisions to the trajectories generated by a policy model. The model exploits trajectories generated by the evaluation policy, while there is not any specific proposal for the explanation logic: Different examples, use different visualizations of transitions’ influences.
4.2.7. Interpretable Reinforcement Learning Using Attention Augmented Agents
A.Mott et al. in (AttAugmentedAgents, ) propose a spatial attention mechanism to visual information in an RL setting. The model enables agents to actively select important, task-relevant information from visual inputs, by sequentially querying and receiving compressed, query-dependent summaries. Actually, the model generates attention maps which can uncover the underlying decision process. Exploiting the attention maps one can understand how the system solves a task, identifying the type of entities that the agent attends (“what”) and the spatial locations where the agent focuses attention (“where”).
Succinctly, the proposed model works as follows: Observations pass through a recurrent vision core network, producing a “keys” and a “values” tensor, to both of which a spatial basis tensor (encoding information about spatial positions) is concatenated. A query network produces a set of query vectors taking as input the state of an LSTM recurrent network (policy network) from the previous time-step. The softmax of the pixel-wise inner product between each query vector and each location in the keys tensor produces an attention map for the query. The attention map is broadcast along the channel dimension, and it is point-wise multiplied with the values tensor. The result is then summed across space to produce an answer vector. This answer is sent to the LSTM to produce the next LSTM state.
The performance of the proposed method is evaluated across a broad range of ATARI levels. Attention maps are used to visualize which parts of the input are attended. Authors conjecture that the top-down nature of the attention (i.e., the fact that the attention map is produced by taking as input the LSTM states) provides a large performance gain compared to equivalent, bottom-up attention-based mechanisms. Comparison of the attention maps to alternate methods for visualizing saliency shows that they allow more comprehensive analysis of the information the agent is using to inform its policy. The agent is able to make use of a combination of “what” and “where” queries to select both regions and entities within the input, depending on the task. Experiments show that the agents are able to learn to focus on key features of the inputs, look ahead along short trajectories, and place tripwires to trigger certain behaviors.
This approach proposes a model addressing the interpretable box design problem for DRL methods applied in visual inputs, where the policy is modeled by an LSTM network. The method focuses on providing global policy explanations. The model exploits LSTM states to produce queries, as well as keys and values tensors produced by the visual input: Thus, we could say that the query network and the vision core serve as the to feed the interpretation component. The interpretable model is a soft attention model and the explanation logic is implemented by visualizing the attention maps, showing the original input frames and super-imposing the attention map for each attention head. In addition, saliency maps, visualizing the relative dominance of “what” and “where” parts of the query are being used, to inspect where the agent focuses attention.
4.2.8. Conservative Q-Improvement
Decision trees that represent policies span the state space and may be larger than required, even if they do express the action value function quite accurately. A. M. Roth et al. in (ConservativeQImprovement, ) aim to provide an answer to the question “what should be the size of such a decision tree” to make it accurate and interpretable? This work is based on the hypothesis that the size should be that, which if increased further it does not increase significantly the estimated discounted future reward of the overall policy.
In doing so, Roth et al. in (ConservativeQImprovement, ) propose the Conservative Q Improvement (CQI) RL algorithm. CQI learns a policy in the form of a decision tree, in any domain with discrete actions and multidimensional state space. In contrast to other RL methods (e.g., (Pyeatt, )) that learn a decision tree policy representation, this approach uses a lookahead approach that predicts which split will produce the largest increase in reward. The method is strictly additive: Initialized with a single leaf node, over time, it creates branches and leaf nodes by replacing existing leaf nodes with a branch node and two child leaf nodes. A branch node represents abstract states and indicates actions to be taken using the function. A leaf node is split (and turned into a branch node) only when the expected discounted future reward of the new policy would increase above a dynamic threshold, moderated by the node visit frequency and the dynamic split threshold.
CQI is evaluated in the RobotNav domain, where a robot must navigate in a 2D environment to reach a goal location, avoiding obstacles. It is shown that the method is able to produce substantially smaller trees compared to the method of (Pyeatt, ), with smaller variance and of greater reward. No explanation logic is proposed in reference (ConservativeQImprovement, ), and this article does not evaluate the method in terms of interpretability or explainability.
CQI follows the interpretable box design paradigm: It aims to learn an interpretable policy model in a direct way towards increasing explainability.
4.2.9. Hierarchical interpretable reinforcement learning
T. Shu et al. in (HierarchicalinterpretableRL, ) propose a framework for efficient multi-task RL, training agents to employ hierarchical policies: These policies support an agent to decide when to use a previously learned policy and when to learn a new skill. Each learned task has a corresponding human language description, thus represented in an interpretable way. Through the hierarchical model proposed, one may i) accumulate tasks progressively from a terminal policy to a top-level policy, and ii) unfold the global policy from top-level to basic actions, using human language descriptions. This approach not only learns the language grounding for visual knowledge and policies, but is also trained to utter human instructions as an explicit explanation of its decisions to humans.
More specifically, training progresses at stages, increasing the set of tasks from to at stage , and the learned policies from for , to for . At stage , the policy is defined by a hierarchical policy that consists of four sub-policies: a base policy for executing previously learned tasks (the policy at stage ), an instruction policy that manages communication between the global policy and the base policy (mapping a task to a base task ), an augmented flat policy which allows the global policy to directly execute actions instead of using a base task, and a switch policy that decides whether the global policy will primarily rely on the base policy or the augmented flat policy. A stochastic temporal grammar (STG) model is trained on the sequence of policy selections (previously learned skill or new skill) and focuses on priorities among tasks: It interacts with the hierarchical policy through modified switch policy and instruction policy. Learning at each stage is a 2-phase curriculum learning task: Policies in each phase are learned with an advantage actor-critic (A2C) off-policy method, assuming that the terminal policy has been trained, with base skill acquisition (learning to use previously learned skills) and novel skill acquisition phases (learning to rely on the augmented flat policy for executing novel tasks).
The approach is evaluated in a two room environment in Minecraft with 24 tasks in total. Experiments aim mainly to show learning efficiency and generalization abilities. Hierarchical plans of several tasks generated by global policies learned by the full model are visualized. However, these visualizations are mainly used to show agent’s learning efficiency, rather than the provision of explanations.
Concluding, this approach follows the interpretable model design paradigm, training interpretable models that are hierarchical plans, where skills and tasks are described in natural language based on human instructions.
4.2.10. LVIN: Imitation learning using the Value Iteration Networks approach
The Logic-based Value Iteration Networks (LVIN) proposed by T. Leech in (LVIN, ) combines imitation learning and learns to represent problems as compact finite state automata (FSA) with human-interpretable logic states. The aim is to facilitate understanding and manipulation of the learned model, towards adding trust and flexibility to robotics applications trained with LVIN.
LVIN produces a policy which describes desired actions in terms of transitions between human-interpretable logic states. In doing so, the planning problem posed by the decision layer is decomposed into two pieces, a high level finite state automaton (FSA) that corresponds to high level sub-goals the robot follows (described in logic) and a low level Markov Decision Process (MDP) describing the motion of the robot in the physical environment. The FSA learns the rules governing the robot’s behavior in terms of interpretable and manipulable logic states, and transitions, called propositions. LVIN comprises separate Value Iteration Networks for each FSA state, where learning the state transitions associated with the high level FSA involves encoding the order of propositions to trigger given the current FSA state, and learning the , and , associated with the low level MDP. LVIN learns the entire model via an iterative update process, similar to Value Iteration Networks, where the update procedure is encoded in a deep NN.
Evaluation aims to show the success rate (i.e., the percentage of correctly completed tasks) of LVIN compared to a CNN trained with a map as input and the desired action as output. The method has been evaluated in a lunchbox packing and in a cabinet checking problem, with no emphasis on evaluating interpretability or explainability.
LVIN addresses the model explanation problem for discrete actions and in a 2D space. The interpretable global model is the high level FSA. In this case, the function provides the “raw” model (deep NN) parameters, i.e., making no interpretation. Since the high level FSA and the low level MDP are trained ”seamlessly” taking as input the same expert demonstrations, learning the interpretable model can be considered as a mimicking process. LVIN does not propose any specific explanation logic for realizing the function . Abstract policies in terms of logical states and transitions explain low level policies, assuming that these abstractions are interpretable by humans. As authors point out “the current LVIN framework … is still somewhat limited by the fact that it requires a hard coded logic oracle for training”.
4.2.11. Interpretable DRL with Linear Model U-Trees
G. Liu et al. in (interpretableDRLwithLMUTs, ) introduce Linear Model U-trees (LMUTs) to approximate NN predictions for DRL agents. An LMUT is learned by an on-line algorithm that is well-suited for an active play setting, where the mimic learner observes an ongoing interaction between the DRL and the environment. The article shows how the interpretable tree structure of an LMUT facilitates analyzing feature influence, extracting rules, and highlighting the super-pixels in image inputs.
U-tree (U-tree, ) (CUT, ) is a classic online RL method which represents a function using a tree structure. In addition to that, continuous U-tree (CUT) (CUT, ) dynamically generate a tree-based discretization of the input signal and estimates state transition probabilities by retaining transitions in every leaf node. To strengthen the generalization ability of these representations, as well as their learning efficiency, Linear Model U-Trees (LMUT) (interpretableDRLwithLMUTs, ) add a linear model to each leaf node. This can be trained either in a batch mode or in an active play setting. LMUT applies Stochastic Gradient Descent to update the linear models, given some memory of recent input data stored on each leaf node. In the active play setting the method exploits transitions in the form of , where the is decided by the DRL model, together with the “soft” , while the is provided by the environment.
LMUT builds an MDP from the interaction data between the environment and a deep model. Compared to a linear Q-function approximator, a Linear Model U-Tree (corresponding to an action) defines an ensemble of linear Q-function models, one per leaf node and one for each state partition cell. Since each Q-value prediction comes from a single linear model, the prediction can be explained by the feature weights of the model.
This approach, has been evaluated towards mimicking a DQN model, and compared against CART, M5 with regression tree, Fast Incremental Model Tree (FIMT) baseline and with Adaptive Filters (FIMT-AF). The evaluation environments include Mountain Car, Cart Pole and Flappy Birds all with discrete actions but with discrete or continuous states. Evaluation concerns approximating the Q values in DQN, reporting on the number of the tree leaves constructed. Furthermore, the sampling efficiency of LMUT is evaluated by computing the correlation and testing error of LMUT, as more transitions are provided. Interestingly, LMUT models are evaluated in terms of the Average Reward Per Episodes by performing the tasks (e.g., playing games) in a direct manner: It is reported that LMUT among other mimic models achieves performance which is closest to the DQN. As far as interpretability is concerned, authors propose (a) a way to measure splitting features’ ”importance” using the total variance reduction of the values, (b) a way to extract and present rules in the form of partition cells, each cell representing a range in specific splitting features, and (c) a way of highlighting pixels (when DRL takes raw pixels as input) while explaining local decisions. However, these options are not evaluated in any setting with human subjects.
Concluding, this is a mimicking approach, where the interpretable model is trained with transitions caused by the interactions of the DRL agent with the environment, also consulting DRL model decisions and values, via the function . The interpretable model represents the agent policy, and as mentioned above, it can be exploited in several ways to provide interpretations. The article does not suggest any particular function implementing an explanation logic.
4.2.12. Policy level explanations for Reinforcement Learning
To address the need of explaining sequences of decisions, N. Topin and M. Veloso in (PolicyLevelExplanations, ) introduce Abstracted Policy Graphs (APGs). APGs are Markov chains of abstract states and are assumed interpretable representations of policies, concisely summarizing them, so that individual decisions can be explained in the context of expected future transitions. The authors propose the APG Gen method for generating APGs for deterministic policies given a learned value function and a set of observed (potentially off-policy) transitions.
An APG is a Markov chain over abstract states where edges are transitions induced by actions from the original MDP. The basic assumption here is that, if the agent’s transitions between grounded states are approximated using a Markov chain between abstract states, then grounded states which are treated similarly (thus, can be treated interchangeably) are readily identified (i.e., in groups) and the agent’s transitions between abstract states can be predicted. Interchangeable ground states grouped into abstract states should have similar future outcomes in terms of rewards, which are assumed known, and the importance of all features for a set of similar states should be low. The importance of a feature is calculated by the Feature Importance Ranking Measure (FIRM), as the variance of the conditional expected score of a state for that feature, with respect to an arbitrary function . This score is the average value of for all states, where that feature takes a specific value .
APG Gen (PolicyLevelExplanations, ) uses states’ similarity based on actions taken under the policy, to repeatedly divide abstract states along important features, and then computes transition probabilities between them. Important features can be trivially recorded for each abstract state, allowing for humans to inspect which features affect agent’s decisions, also providing a summary of an abstract state and of the corresponding ground states.
APG is evaluated in an abstraction of a production task for a multi-component item using a number of manufacturing steps, in deterministic and stochastic settings. Policies are computed using value iteration. In providing local explanations, APG is evaluated in terms of the features’ prediction accuracy (important vs. not important) as a function of the portion of states sampled: This provides a view on how APG generalizes on features’ importance, beyond the states seen. Additionally, n-hop action predictions are evaluated to measure the error caused by assuming that only single actions are performed for a transition between abstract states. This is measured by the action prediction agreement (predicted vs actual) for increasing time horizon. Finally, the size of “explanations” are measured by the number of abstract states (nodes in APG) against the number of grounded states in the MDP (reported sub-linear). No evaluation with human subjects in simulated or real-world settings are reported.
Concluding, APG Gen is a mimicking approach, with the objective of providing an interpretable model of a learned policy, considering a trained policy model. That model is realized by an APG. APG Gen takes as input transition samples from the learned policy, the value function, as well as features’ importance. APG is the interpretable model, and local explanations may be provided consisting of the features that are important for a specific state. Reference (PolicyLevelExplanations, ) does not propose an explanation function for providing interpretations’ surface representations.
4.2.13. PIRL: Programatically interpretable reinforcement learning
PIRL (Programmatically Interpretable Reinforcement Learning) is an RL framework (PIRL, ) designed to generate agent policies that are represented using a high-level, domain-specific programming language. Such programmatic policies are not as performant as those from DRL methods, but are interpretable and being amenable to verification by symbolic methods. A. Verma et al. in (PIRL, ) propose the Neurally Directed Program Search (NDPS) method for solving the challenging non-smooth optimization problem of finding a programmatic policy with maximal reward, minimizing the distance from an “oracle” DRL policy.
Specifically, considering a Partially Observable Markov Decision Process (POMDP) setting, a functional language, and given a policy sketch that syntactically defines a set of programmatic policies, the objective is to find a program in this set with maximal long-term reward. To restrict the search, the NDPS algorithm performs directed policy search guided by an oracle policy learned by a DRL method, via local search. NDPS measures the “distance” between the outputs of a programmatic policy and the outputs of the oracle policy, given a set of “interesting” histories which are enriched with additional histories generated by the programmatic policy (a technique that in (PIRL, ) is called “input augmentation”). Further optimization improves the programmatic policies found (e.g., being shorter).
PIRL is evaluated on the task of learning to drive a simulated car in the TORCS car-racing environment, using as oracle a policy computed by the Deep Deterministic Policy Gradient (DDPG) algorithm: NDPS policies are compared to human-readable policies, also showing that PIRL policies can have smoother trajectories, and better generalization abilities compared to the policies discovered by DRL. No evaluation of the explainability or interpretability of PIRL is provided, with or without human subjects.
Concluding, PIRL with NDPS is a mimicking process, exploiting agents’ interactions with the environment to find optimal interpretable programmatic policies that are presented as functional programs.
4.2.14. Understanding Decisions by Interpretable Policy Learning
INTERPOLE is an imitation learning method that jointly models the agent’s policy (belief-action mapping) and belief-update process, towards interpreting demonstrated behaviour. As such, it can be used to model any “expert” demonstrated policy, without assuming unbiasedness of agent’s beliefs or optimality of policies. In so doing, A. Hüyük et al. in (ExplainingByImitation, ) propose an inherently interpretable imitation learning method, locating factors that contribute to individual decisions, in a language that is readily understood by domain experts, accommodating partial observability and operating offline.
INTERPOLE aggregates sequential observations through agent’s subjective belief-update process (decision dynamics) and probabilistic belief-action mapping for determining actions (decision boundaries). The objective is to model the most likely parameterizations , drawn from some prior , i.e. the likelihood of with respect to a given set of demonstrated observation trajectories and unobserved state trajectories : is the state transition probability, is observation’s probability after a transition to a new state, is the probability of being in state in , the smoothness of transitions between decision boundaries is tuned by the temperature factor , and are actions’ mean vectors inducing decision boundaries over the belief space. Parameters are jointly determined by means of an expectation-minimization (EM)-like algorithm for maximizing , given that is not known. Estimating jointly the parameters , INTERPOLE learns agent’s decision and belief dynamics (not the environment dynamics) offering the most plausible explanation of how the agent reasons.
INTERPOLE has been evaluated using simulated and real world demonstrated trajectories for disease diagnosis, focusing on interpretability, accuracy of learned policies, and subjectivity towards recovering interpretations of behaviour even if the agent is biased. For interpretability, INTERPOLE has been evaluated by nine clinicians focusing on the utility of representing possibly subjective action-belief trajectories instead of raw action-observation trajectories, and on understanding policies by means of suboptimal decision boundaries, instead of using a reward function: The majority of clinicians preferred the explanations offered by INTERPOLE. Regarding accuracy, INTERPOLE is evaluated in modelling the belief-update process and the belief-action mapping (policy), learning high quality models driving agent’s subjective reasoning.
INTERPOLE (ExplainingByImitation, ) addresses the policy explanation problem, offering a method for imitating agents’ possibly suboptimal policies without assuming unbiasedness beliefs and objective reasoning. The interpretable model comprises decision dynamics and decision boundaries and is specified by without assuming a fully observable setting, given a set of demonstrated action-observation trajectories . The explanation logic visualizes a belief simplex in low dimensional spaces, showing decision boundaries and belief trajectories.
4.2.15. Automatic discovery of interpretable planning strategies
With the aim to assist human experts in decision making, J. Skirzyński et al. in (InterpretablePlanningStrategies, ) introduce the Adaptive Imitation-Interpretation (AI-Interpret) algorithm for transforming policies into sets of interpretable descriptions by means of decision rules that have a disjunctive normal form (logical program policies), presented as flawcharts.
Specifically, given a set of demonstrated trajectories produced by the original policy, a domain specific language specifying the predicates for constructing the logical program policies, a parameter defining the threshold at which the discovered policy (represented by a logical formula) is at least better than the original policy, and the maximum rule length () of the logical formula, the objective is to find the logical formula that maximizes the number of demonstrations in that it can imitate w.r.t. and . AI-Interpret uses the predicates to separate the set of demonstrations into clusters. Clusters (whose number is determined automatically) are evaluated by means of a heuristic value that shows the similarity of trajectories in the cluster and how representative the cluster is. Doing so, the algorithm can consider increasingly smaller sets of demonstrations (disregarding less important clusters) and employ Logical Program Policies (LPP) (LPP, ) in a structured search searching for the simplest logical formula w.r.t. and that imitates most of the demonstrated trajectories. Finally, the formulae are transformed into decision trees, and then visualized as flowcharts that people can follow to execute the strategy.
AI-Interpret is evaluated in three types of challenging planning tasks. The central question is whether discovered flawcharts can support decision-makers in the process of planning. To achieve that, authors in (InterpretablePlanningStrategies, ) perform one large behavioral experiment for each of the three types of tasks and a fourth experiment in which they evaluate whether AI-Interpret improves human decision-making against conventional training (i.e., by providing feedback). Results show that the proposed approach (a) allows people to largely understand the automatically discovered strategies and use them to make better decisions, and (b) is more effective than conventional training, improving human decision-making.
Overall, AI-Interpret (InterpretablePlanningStrategies, ) addresses the policy explanation problem, discovering policies described by means of logical programs. This is a mimicking approach, as the input it requires are demonstrations from the closed-box RL policy model. The interpretable model is the LPP discovered, and the explanation logic transforms LPP to flowcharts.
4.3. Solving the objectives explanation problem
4.3.1. Enabling robots to communicate their objectives
S. H. Huang et al. in (CommunicateObjectives, ) conjecture that, end-users, to understand and predict what the agent will do, need to have an accurate mental model of the agent’s objective function. To reduce the time needed for this otherwise time-lengthy process, authors propose to model how people infer objectives from observed behavior, in order agents to demonstrate behaviors that are maximally informative.
Following an algorithmic teaching approach, authors in (CommunicateObjectives, ) define candidate models of human inference regarding robot’s objective function, given robot’s optimal behavior in specific environments: These demonstrations increase the probability of humans to infer the correct objective function. The objective function is assumed to be a linear combination of (possibly complex functions on) environment features in , weighted by parameters . Features are supposed to be known by humans (although not always the case), who need to learn the true reward parameters. This involves searching for a sequence of environments , such that when the person observes the optimal trajectories in these environments, their updated belief places maximum probability on the correct reward parameters. The challenges tackled to solve this optimization problem are as follows: (a) Humans are likely to be approximate in their inference, while (b) the agent needs to incorporate a model of this approximate inference, (c) encouraging full coverage of all possible strategies that the agent is capable of adopting. Updates of humans’ beliefs for the reward parameters can be modelled either via an inverse reinforcement learning technique (exact method), or via Bayesian inference (approximate method), modelling how probable humans consider trajectories to be, given their beliefs on the reward function. To address the coverage challenge, a new term is added to the formulation of the optimization problem, ensuring that no extra trajectory examples are selected unless these examples increase significantly the probability of inferring the true reward parameters. Extra examples are chosen to be the best with respect to the best approximate model.
Experimental results evaluate how approximate and exact inference models perform in a driving simulation environment. The aim is the system to provide to participants examples of how the car drives, with the goal of being able to anticipate how it will drive when they ride in it. The analysis of models is done with ideal users and with human subjects. Analysis shows that while not just any approximate-inference model is better than the exact one, choosing a suitable approximate model helps significantly. The user study on coverage concludes that coverage with the right model of approximate inference have a significant advantage over RL models assuming exact inference users.
Concluding, reference (CommunicateObjectives, ) provides methods that address the objective model explanation problem, focusing on teaching the agent’s reward function true parameters via examples. The proposed approach explicitly incorporates a model of the human beliefs on reward parameters and aims to provide a view on the true objective function. While the interpretable model can be considered to be the linear objective function, no explanation logic is provided here.
4.3.2. Explaining agent behaviour through intended outcome
H. Yau, C. Russell and S. Hadfield in (WhatDidYouThink, ), aim to interpret decisions by describing agents’ intended outcome. This method applies to RL methods that estimate values off/on policy, given discrete states and actions and a deterministic MDP. The approach is demonstrated to off-policy -learning and aims to recover the future events that impact the -values learned.
This approach learns a belief map of discounted expected sum of state visitation H. H is updated when a Q-value is updated, as follows: , where denotes an indicator function with 1 if the agent is at and 0 otherwise.
Given a deterministic reward function , the belief map H is consistent with if . Authors prove that if H and are zero-initialized, then H and are consistent for all iterations of the algorithm in the tabular case. For deep Q-learning methods, there is no guarantee for consistency of H with Q.
The proposed approach is evaluated in domains with deterministic or stochastic reward functions and with continuous or discrete state-action space. In environments with stochastic rewards and continuous states, new discrete states are added, or states are discretized. The method is demonstrated to work effectively for tabular and deep Q-learning agents.
This is a generic approach that addresses the objective model explanation problem, providing projections of future intended trajectories of agents. The interpretable model is provided by the belief maps, and explanations are provided by means of belief maps visualizations showing the value of each intended future state-action pair. The method is limited to low-dimensional state-action spaces.
4.3.3. Distal Explanations for explainable reinforcement learning
P. Madumal et al. in (DistalExplanations, ) emphasize on the role of opportunity chains for explaining agents’ behaviour, and introduce a distal explanation model that can generate opportunity chains as explanations for model-free RL agents. An opportunity chain takes the form of “A enables B and B causes C”, where B is the distal event or action, and can inform the explainee about long term dependencies between events, when certain events enable others.
Reference (DistalExplanations, ) proposes a distal explanation model that learns opportunity chains using a recurrent neural network (RNN). As distal explanations by themselves would not make a complete explanation, this approach uses action influence models (CausalLens, ) that approximate the causal model of the environment relative to actions taken by an agent, to get the agent’s goals. It further improves upon action influence models (specifying the effects of actions in features) using decision trees instead of structural equations used in (CausalLens, ) (presented in Section 4.4.7) to represent the agent’s policy. The decision tree policy model is trained concurrently with the RL policy model, assuming a model-free algorithm and exploiting state-action samples using an experience replay buffer. The tree is constrained to have a max number of leaves, equal to the agent actions. The same dataset of samples exploited for learning the decision tree model is exploited by a prediction RNN to approximate distal actions and their cumulative rewards.
The accuracy of distal prediction and counterfactuals is evaluated in six benchmark domains using different model-free RL algorithms. Benchmarks have a mix of complexity levels and causal graph sizes (number of actions and state variables). Results show that the distal explanation model is robust and accurate across different environments and algorithms. Furthermore, experiments with humans using RL agents report on task prediction (i.e., understanding of the agent) and subjective explanation satisfaction. Agents are trained to solve 1) an adversarial task; 2) a search and rescue task; and 3) a human-AI collaborative build task; all in Starcraft II. The results obtained for explanation quality show that distal explanations perform substantially better than baselines in human-agent collaborative tasks.
Concluding, this is a mimicking process, where the interpretable models (decision tree and opportunity chains) are trained by exploiting RL agent’s interactions with the environment. The interpretable models comprise a policy model using a decision tree, and a causal model representing the objectives (), as well as the importance of specific responses using opportunity chains. The explanation logic exploits models to produce explanations in natural language.
4.4. Solving the outcome explanation problem
4.4.1. Reward Decomposition
The approach proposed by Z. Juozapaitis et al. in (RewardDecomposition, ) decomposes rewards into sums of semantically meaningful reward types, so that actions can be compared in terms of trade-offs among the types, with a focus on explainability. In particular, this work introduces the concept of minimum sufficient explanations for compactly explaining why one action is preferred over another in terms of the reward types.
Specifically, assuming that the MDP formulation incorporates a set of reward components/types and defining a vector-valued reward function with a component for each , , the agent objective is to optimize the overall (mixed) reward function . This vector-valued reward allows for defining a vector-valued -function, with action-value components for rewards of type . The overall Q-function is defined to be . This results in a straight-forward heuristic tabular algorithm updating values in a standard way for each of the components, called HRA. The definitions are extended to DRL settings, where each is parameterized by means of . This results into DRL algorithms for decomposed reward (dr), drDQN and drD-SARSA.
To gain insight into why an agent prefers action over in state , i.e., , this approach uses the difference explanation (RDX) as the difference of the decomposed -vectors. RDX is exploited to find the minimal sufficient explanations and for each pair of actions and state, indicating the “critical” positive and negative reasons (in terms of reward types), respectively for the preference.
Case studies in two environments illustrate the potential of visual explanations in the hands of an RL practitioner. Authors show how decomposed reward methods support spotting certain “bugs” in the agent’s action values and even help identify a more subtle issue related to the interaction of the gradient optimizer and the DRL loop.
Concluding, this approach addresses the outcome explanation problem, focusing explaning decisions on actions against other actions locally, i.e., in specific states, exploiting reward types. The interpretable model takes the form of tuples per state and applicable action couples, exploiting (decomposed) action values provided by . The explanation logic is implemented by means of bar charts on features.
The reward decomposition approach has been proposed as the basis for a set of decision-making tasks in (StrategicTasks, ), in environments that naturally provide multiple reward signals: The goal is to enhance explainability by providing high-level abstractions for sequential tasks.
4.4.2. Autonomous Policy Explanation
B. Hayes and J. A. Shah in (AutonomousPolicyExplanation, ) aim to enable an autonomous agent to reason over and answer questions about its underlying control logic , independent of its internal representation. First, it learns a domain model of the agent’s operating environment from real or simulated demonstrations, exploiting programmer specified code annotations for actions, and variables (state parameters/attributes) affected by actions. Within this learned domain model, this approach uses statistics computed over data extracted from continued observations to construct a behavioral model that approximates the agent’s control logic. The article shows applicability of the proposed approach in tabular Q-learning, deep Q-learning and hand-coded controller logic.
To explain the policy given a query , this work provides an interpretation function to provide a natural language response . is constrained to state parameters and annotated functions. is constrained to a template-based approach using natural language.
The interpretation function is defined to be , where, resolves the question identified by to a set of problem states (context), and summarizes attributes of problem states into a comprehensive representation. The realizes the function , providing a representation of relevant parameters to resolve a question. Depending on the question type (see below), this may take the form of searching and gathering state parameters’ from states in a region, or from similar states given a specific state, or from near-by states. Questions concern when an agent will behave in a specific way, how it will behave under specific conditions, and deviations from the expected behavior. Question-answering is template driven using a response resolution algorithm per query type. Depending on the question type, provides an explanation for a specific response (i.e., locally) at the granularity of policy actions, explaining the control logic in specific contexts (i.e., sets of states). The explanation logic is implemented by the function and is realized by means of Boolean classifiers using communicable predicates, which are similar to STRIPS-style planning predicates, associated with natural language descriptions.
This work allows human collaborators to shape their expectations on agents’ behaviour without requiring a full understanding of an agent’s logic, and facilitates the debugging of aberrant behaviors by explaining differences between the conditions under which particular actions occur. The article demonstrates the applicability of the proposed method within three representative robotics domains with discrete and continuous state parameters, in single/multi agent settings, using different types of controllers. However, experiments focus on the applicability of the method, rather than on the accuracy/fidelity and conciseness of the explanations provided in any specific human-robot collaborative setting.
4.4.3. Transparency Communication for Reinforcement Learning
N. Wang et al. in (TransparencyCommunicationforRL, ), as a continuation of previous work on static models (TransparencyCommunicationforRL, ), discuss the design of model-based and model-free RL for robots in a human-robot simulation testbed. The goal is to provide communications for the robot’s decision making update process towards interpretability and repairing trust. This is achieved by interpreting components of the robot’s decision making and learning process, assuming that the robot bases its decisions on a POMDP model.
Considering a dynamic POMDP and assuming a model-based RL that updates the POMDP model, the proposed method interprets a static POMDP after each update, informing the teammate about changes. This work focuses on updating the observations model, i.e., the probability for the robot to get any observation. Additionally, assuming a model-free RL, values account for changes in a possible model, but with no explicit representation of these changes and model. To address this last issue, authors propose finding a potential POMDP that is consistent with the policy arrived at by the model-free RL. To do that, they compute the optimal policy for that POMDP exploiting the assumption of a piecewise-linear model (PieceWiseLinearFunctions, ), representing that policy by means of a decision tree, which is compared to the RL policy. If it matches, this POMDP is used to provide explanations.
This article does not report on the efficacy of explanations with human-subjects. As authors also point out, this must be done with different permutations of explanations’ aspects, also providing insight into what aspects should be included in the explanations and how to present them.
Concluding, this approach provides interpretations of POMDP components, with no access to the learning process itself. We can consider that functions provide updates on agents’ beliefs, transition models, or updates in the values, while the interpretation process updates the existing POMDP, or searches for a potential matching POMDP, which serves as the interpretable model. Function provides explanations on POMDP components’ updates using natural language static templates. In addition to revealing information about individual POMDP components, templates can be created informing about the overall model and how the learning arrived at that model, also leveraging the interpretable policy models generated by exploiting the piecewise-linear assumption.
4.4.4. Transparency & Explanation in DRL
R. Iyer et al. in (TransparencyandExplanationInDRL, ) report on the interpretability of DRL Networks (DRLN) in visual tasks, proposing the Object-sensitive DRL (O-DRL) method that (a) incorporates explicit object recognition processing into DRL models, (b) forms the basis for the development of object saliency maps, and (c) can be incorporated in any existing DRL framework such as DQN and A3C. The goal here is to produce intelligible visualizations of agents’ state and behaviour, focusing on objects saliency, i.e., on the influence of objects features in agents’ decisions.
The proposed O-DRL approach uses a deep neural network (CNN) that takes as input object channels and original images to predict -values. Object channels represent types of objects recognized in an image and incorporate objects’ features. To compute object saliency, this proposal masks the object and computes the values on the new (perturbed) state, and the difference of new values to the values in the original state.
Authors report on experiments with humans, regarding the behaviour of a Pacman agent. The goals of the experiments were to (a) test whether object saliency maps contain enough information to allow humans to match them with corresponding game scenarios, (b) test whether participants could use object saliency maps to generate reasonable explanations of the behavior of the Pacman and (c) test whether object saliency maps allow participants to correctly predict the agent’s next action. Experiments include (a) a matching task, where participants match saliency maps to agents’ behaviour; and (b) a movement prediction task from video clips with or without saliency maps. Results show that object saliency maps can be linked to corresponding game scenarios by participants (thus, they do provide a basis for visual explanations) but, as results from the prediction task suggest, the exact use of salient maps must be further investigated in conjunction to providing rich contextual information about each situation.
This is an approach that addresses the outcome explanation problem in visual tasks, focusing on the influence of objects in agents’ decisions in specific states. The proposed interpretation method is realized by the O-DRL method that exploits values, provided by function to rank importance of features (image pixels or objects). Interpretable models are realized by pixels’ and objects’ salience. The explanation function realizes surface representations of interpretations using visual means.
4.4.5. Understanding agent actions using specific and relevant feature attribution
Recognizing that perturbation-based approaches for RL, as that proposed by S. Greydanus et al. in (VisualizingAtariAgents, ) and the O-DRL approach (TransparencyandExplanationInDRL, ) in 4.4.4, tend to produce saliency maps that are not specific to the action of interest, N. Puri et al. in (ExplainYourMove, ) propose SARFA to generate saliency maps that focus on explaining the specific action decided, balancing between specificity and relevance.
SAFRA is a pertrubation-based approach for generating saliency maps focusing on capturing the impact of pertrubation only on the value of the action to be explained (specificity), and downweighting features that alter the expected rewards of actions other than the action explained (relevance): Specificity is measured by capturing the relative change to the value for the action to be explained w.r.t. other actions, and relevance by considering how the relative expected reward of taking some actions (other than the one explained) changes with a pertrubation. The harmonic mean of these two measures produces the salience of state features.
SAFRA in (ExplainYourMove, ) has been evaluated qualitatively in games (chess, Attari games and Go), showing that in comparison to O-DRL (TransparencyandExplanationInDRL, ) and the approach in (VisualizingAtariAgents, ) produces more focused saliency maps. In addition, human studies on problem-solving chess puzzles show that saliency maps produced by SAFRA are less confusing than those produced by (TransparencyandExplanationInDRL, ) and (VisualizingAtariAgents, ). SAFRA is compared to the other approaches using a chess saliency dataset produced by human experts, showing superiority in identifying salient features and robustness in pertrubations.
Overall, SAFRA (ExplainYourMove, ) addresses the outcome explanation problem, focusing on visual tasks and aiming to identify state features whose salience is specific and relevant to the decided action. SAFRA exploits values, provided by and interpretation is provided by the salience of state features. The explanation is surfaced using visual means, showing the salience of features.
Regarding the power of salient maps to assess the degree to which hypotheses about features of the learned policy correspond to the semantics of RL settings, A. Atrey et al. in (ExploratoryNotExplanatory, ) suggest that salience maps should be used as an exploratory rather than as an explanatory tool and propose a methodology grounded in counterfactual reasoning, focusing on issues of subjectivity, unfalsiability and cognitive biases on mapping salient features to semantic concepts and behaviours. In addition to that, saliency maps have been used also in combination with policy explanation methods in (LocalAndGlobal, ), investigating their role when put in the context of agents’ strategies (discussed in Section 4.2.1).
4.4.6. Autonomous Self Explanation of Interactive Reinforcement Learning
Y. Fukuchi et al. in (AutonomousSelfExplanation, ) propose Instruction-based Behavior Explanation (IBE) to explain agent’s future behavior within a sufficient temporal extent, while the policy is being learned in an interactive RL setting. In IBE, an agent exploits the instructions given by a human expert (which may not indicate the actions to be performed) to explain its own behavior: This happens under the assumption that when the agent receives more rewards it is likely that it followed the instructions given.
IBE consists of two steps: (i) estimating the target of the agent’s actions by simulation: the target is the change in the environment state after a fixed temporal extent for explanations specified by agents’ actions in steps, and (ii) acquiring a mapping from the target to the instructions, through a k-means classifier, in order to explain the action target based on the instruction signal given by a human expert. Authors report on experiments with humans in a game setting, using two DQN agents, each at a different training stage. The most important finding is that the environment changes must be chosen deliberately for assigning an explanation signal: The parameter that defines the number of agent’s actions to a target state can not be fixed, while predictions in settings of high-complexity must be accurate.
IBE (AutonomousSelfExplanation, ) provides a local interpretation method, interpreting the decision of the agent in specific states, given a target state in a constant temporal extent of time steps. The interpretation function is the IBE method itself, exploiting agents’ target and instructions provided by experts. Thus, the RL method is treated as a dark box, where the function fetches instructions provided by experts, provided as representations of agent’s behaviour. The classification of assessed targets to instructions provides a local interpretable model. The explanation logic is context and domain dependent: examples of visualisations are provided in (AutonomousSelfExplanation, ). As stated in (AutonomousSelfExplanation, ), prediction of changes is challenging in complex settings, and the temporal extent may need to vary in different contexts, even for the same agent.
4.4.7. Explainable reinforcement learning through causal lens
P. Madumal et al. in (CausalLens, ) aim to generate explanations using causal chains from action influence graphs, for model-free RL agents. This is accomplished by encoding causal relationships between variables of interest and learning a structural causal model (StructuralCausalModel, ) during RL. Action influence models approximate the causal model of the environment relative to actions taken by an agent. This approach uses causal models to generate contrastive explanations for why and why not questions, and formulates the minimally complete explanations, and minimally complete contrastive types of explanations to be provided from an action influence model.
A critical part of this approach is learning the structural equations, i.e., the quantitative influences of actions on variables: Assuming that a DAG specifying causal direction between variables is given, structural equations are approximated as multivariate regression models during the training phase of the RL agent, exploiting samples of agent interaction with the environment via an experience reply. Therefore, this is a mimicking process, where interpretable models are structural equations specifying the effects of actions in features, the causes of rewards, and the states resulting in rewards. These local models provide local response and local objectives’ explanations. The explanation logic , provides explanations to the two types of questions: “why” and “why not”.
This approach has been computationally evaluated on six RL benchmarks domains using six different model-free RL algorithms with discrete actions. Results indicate that the proposed models are robust and accurate enough to perform task prediction with a negligible performance impact. A human study (120 participants) evaluated the method in task prediction, explanation satisfaction, and trust, showing that the proposed model performs better than the tested baseline, but its impact on trust is not statistically significant.
4.4.8. Self-supervised discovery of causal features
S. Wenjie et al. in reference (Self-SupervisedDiscoveryOfCausalFeatures, ) propose a self-supervised interpretable framework (SSINet) to explain a trained policy model of vision-based RL agents to locate fine-grained causal features towards gathering state-specific and task-relevant evidence for the agent’s decisions.
Specifically, given a trained policy model, the proposed framework learns an explanation model that predicts an attention mask. The basic attention patterns learned are exploited for identifying the relative importance of features and analysing failure cases. If the generated actions are consistent when the policy takes as input either the state or the attention-overlaid state, the features highlighted by the proposed method are considered to be task-relevant and constitute evidence for the agent’s decisions. The agent in (Self-SupervisedDiscoveryOfCausalFeatures, ) is pre-trained using model-free RL algorithms including proximal policy optimization (PPO), soft actor-critic (SAC) and twin delayed deep deterministic policy gradient (TD3).
More technically, the SSINet uses a U-Net architecture, providing “masked” states through a feature extractor, a mask decoder and a sigmoid non-linearity. SSINet uses the same feature extractor with the pretrained agent policy model, thus SSINet trains only the mask decoder. SSINet is trained using a dataset of state-action pairs generated by the policy. Training is “driven” by a mask loss function aiming to satisfy two desiderata: Maximum behaviour resemblance, i.e., the agent’s behaviour using masked states must be as consistent as possible with that when using original states, and minimum region retaining, i.e., attentions must attend as little information as possible.
SSINet has been evaluated on several Atari 2600 games, as well as on Duckietown, a self-driving car simulator environment. Empirical results verify the effectiveness of the proposed method, and demonstrate that the SSINet produces high-resolution and sharp attention masks to highlight task-relevant information. Special emphasis has been given to explaining why the agent performs well or badly quantitatively, by introducing evaluation metrics that assess the effectiveness of attention masks, in multiple RL algorithms and actor architectures.
In conclusion, the SSINet proposed in (Self-SupervisedDiscoveryOfCausalFeatures, ) is a self-supervised machine learning process for vision-based DRL agents. Although it exploits samples from a trained policy model, it does not aim to produce an interpretable model that mimicks or that distills the knowledge acquired in the policy model: It rather addresses the outcome explanation problem, providing a local model that identifies salient state features, which are further presented via appropriate visual masking techniques realized by , applied in the original state.
4.4.9. Distilling DRL in Soft Decision Trees
Y. Coppens et al. in (DistillingDRLInSDTs, ) illustrate how Soft Decision Trees (SDT) (DistillingNNsInSDTs, ) can be used as interpretable policy models. SDT are hybrid classification models of binary trees of predetermined depth, and NNs. Each branching node represents a hierarchical filter that influences the classification of input data. The leaf nodes learn softmax distributions over possible classes using model parameters , while the overall loss is minimized using mini-batch gradient decent optimization.
Using state action pairs generated from the policy learned using the actor-critic Proximal Policy Optimization (PPO) algorithm, the soft decision tree is trained to classify states to actions. The applicability of the method is evaluated on the Mario AI benchmark: Authors show the fidelity with which SDTs can provide interpretations, and elaborate on the performance of these models when one would consider to use SDTs directly for task execution.
This is a mimicking process (as authors in reference (DistillingDRLInSDTs, ) also state), despite the article title that characterizes the approach as a distillation process. The model is the soft decision tree, which provides local only interpretations, classifying states to agent actions. Despite the possibility of examining the learned filters along any decision tree path from the root to leaf, it is questionable whether the soft decision tree offers a fully interpretable model, given the parameterized leaf nodes and the decision tree filtering nodes leading to an action distribution. As noted in the article, since the distributions in the leafs need to generalize and provide an action strategy for multiple state samples, it is to be expected that the resulting action distributions in the soft decision trees can differ from the NN output of the PPO agent for individual input samples: This results in low fidelity in mimicking the PPO policy. The interesting aspect of this work is that, bigger decision trees, offer greater agent rewards. However this comes with the cost of reduced interpretability. The explanation logic provides surface representations using hitmaps, exploiting the SDT model in spatial settings. The use of SDTs in non-spatial settings is something that worth further investigation.
4.4.10. Memory-based explainable reinforcement learning
F. Cruz, R. Dazeley and P. Vamplew in (MemoryBAsedXRL, ) propose a memory-based explainable reinforcement learning (MXRL) approach according to which the agent exploits an episodic memory to infer the probability of success and the number of transactions to reach the goal state. Specifically, the agent exploits past interactions with the environment, and by “introspection” can compare the probability of choosing an action against the probability of being successful, thus providing explanations in terms of the necessity to complete an intended task. The interpretable model aims to explain the agent’s decision for an action at a specific state, based on past experience, exploiting transitions in successful runs. It is assumed that all transitions have been recorded. This limits the method applicability to small, discrete state - action spaces, such as the grid world in which experiments have been conducted.
In essence, this is a distillation process where the agent distills knowledge regarding the reasons that the RL agent favors specific actions for specific states, through past experience gathered in an “interpretation” that applies at the episodic memory. The explanation logic is provided via hitmaps and diagrams, showing probabilities for choosing actions in states, and probabilities to succeed in reaching a goal state.
4.4.11. Enhancing explainability of DRL through selective LRP
As noted in (Miller, ), people usually prefer explanations that focus on selected, specific evidence, instead of showing every possible cause of a decision. Based on this insight, authors in (LRP-argmax, ) aim to adjust the layer-wise relevance propagation (LRP) (LayerWiseRelPropagation, ) saliency map approach to focus on the parts of the input that are most relevant for the decision-making process of a DRL agent. Their contribution is twofold: (a) Their adjustment uses an argmax function to follow only the most contributing neurons, filtering out the most relevant information; (b) they adjust LRP to dueling DQN convolutional layers.
LRP describes a concept that can be applied to any classifier if that fulfills the following two requirements. First, it has to be decomposable into several layers of computation where each layer can be modeled as a vector of real-valued functions. Second, the first layer has to be the input of the classifier and the last layer has to be the real-valued prediction of the classifier . Any DRL agent fulfills those requirements, considering actions decided as the output: In this case LRP can be used to address the outcome explanation problem. According to the LRP concept, relevance values are assigned to each computational unit of each layer in such a way that measures the local contribution of the unit to the prediction . The calculation of values follow the LRP concept, if it sets the relevance value of the output unit to be the prediction and calculates all other relevance values by defining for messages , such that . Relevance values can be computed in a backward pass, starting from the output layer, towards the input layer. This concept has been applied to the dueling DQN architecture in (LRP-argmax, ), specifying the backward pass through the fully connected and the convolution layers. However to find the most relevant input neurons contributing to a specific decision, authors propose the argmax rule, defining the messages as follows:
where are model weights and the inputs.
This approach has been evaluated on three Atari 2600 games to verify that the saliency maps generated are more selective than the ones created by alternative LRP methods, while still including the information expected from visual explanations: This can be beneficial to the transparency of DRL methods.
Overall, this is an approach that addresses the outcome explanation problem, aiming to provide the most salient features towards specific DRL agents’ decisions. The model constructed concerns the local contributions (relevance values) of inputs and neurons to the decision made, while explanations are provided visually by means of inputs’ saliency maps.
5. Concluding Remarks
Explainability of closed-box methods that provide agents with the capacity to act autonomously in the real world, and particularly of DRL methods, is rather challenging, albeit emerging: DRL, being successful to prescribe actions and courses of actions in complex settings according to the evolution of the environment, may comprise a number of closed-boxes, including models of the environment, of the agent objectives and the agent policy, as well as models that fit states’ or state-action pairs’ values. Based on the general blueprint of DRL methods, in this article we formulate specific explanation problems: (a) The policy and objectives model explanation problem, aiming to provide interpretation of the overall - global - control logic behind specific agent’s responses prescribed by the agent’s policy and motivated by the agent’s objectives, in coarse granularity and large scales; (b) the responses and local objectives explanation problem that aims to interpret agents’ timely - local - responses in specific states, also explaining the correlation between environment’s state features with agents’ objectives in fine granularity and small scale; and (c) the model inspection problem, providing elements and properties of any DRL individual model.
The article identifies specific paradigms that have been proposed for implementing XDRL methods, proposing a distinction between the distillation and mimicking approaches. It proceeds to provide a review of state of the art methods for XDRL methods, addressing the needs of human operators - i.e., of those that take the actual and critical decisions in solving real-world problems. For each state of the art proposal the article describes the overall objectives and the specific explanation problem that is being addressed, the method used for providing interpretations, the interpretable models constructed according to the DRL explainability paradigm followed, the explanation logic and the surface representations of explanations.
What can follow as a general and straightforward remark from this review, is that while there is a lot of work and progress towards solving XDRL problems, it is very early to conclusions regarding DRL explainability and transparency: We need to understand the possibilities to their full extent in all aspects across the pipeline, and in all different dimensions regarding the explainability desiderata. Indeed, there are many questions that remain unanswered, which present major challenges. Subsequent paragraphs try to point out the major and important ones, according to our view.
Starting from the questions that are stated in the introductory part of this article, there is not any work that identifies and provides evidence about the qualities of “good” and objective explanations for RL agents. There is some evidence about the qualities of explanations from other scientific fields (e.g., for an overview in such evidence in social sciences one may start from (Lipton, )), but we need to bridge these to the effectiveness of RL agents’ explanations (for any of the explanation problems stated), also considering the characteristics of the real-world contexts in which agents are deployed. Specifically, while many approaches make specific proposals for interpretation and presentation of explanations, these are made in a fragmentary and ad-hoc manner (e.g., exploiting few natural language templates, or hitmaps with specific characteristics) providing mixed-results or no evidence on the effectiveness of these explanations (maybe in combination with other modalities) in real-world settings. For vision-based RL, for instance, a feasible explanation approach is to learn t-Distributed Stochastic Neighbor Embedding (t-SNE) maps, while, there are works visualizing important features for RL agent’s decisions using saliency maps: The role of these maps to the effectiveness of providing local explanations needs further investigation, also in combination with other methods that provide contextual and more broad perspectives of agent behaviour.
Interpretable machine learning presents many challenges (e.g. some of them are discussed in (InterpretableML, )), which are inherited to constituent models of deep reinforcement learning methods. However, RL agents that solve sequential problems need to provide explanations on sequences of decisions and exploit the constituent models’ interpretations in intertwined ways. Bellow we present a list of challenges that we consider important to report essential progress in explainable deep reinforcement learning in particular:
- Build a toolbox for explaining DRL: We need as general as possible and advanced tools for the surface representation of explanations from DRL agents: These must identify generic types of information to exploit and must include appropriate configuration functionality to provide information using multiple modalities, allowing users to explore explanations at different levels of detail, granularity and scale. Such a toolbox would provide explanations’ content exploiting specific types of information and transforming information (e.g., via projections, filtering, aggregation and extraction of features, etc.) in comprehensive ways, for any combination of problem dimensions.
- Be as comprehensive as needed: We need to realize approaches that interpret and explain the behaviour of DRL agents in as much as possible comprehensive ways: Such approaches should address combinations of the explanation problems stated, in a joined way, supporting explaining the overall policy and objectives model, specific responses, indicating features importance in general, refining features’ ranking in specific cases and inspecting the consequences, indicating parts of the state-action space of special interest at different levels of granularity and scale. This will allow synthesizing comprehensive explanations for humans to understand agents’ decision making, tailored to specific context of agents’ deployment, with respect to users’ requirements, constraints and cognitive aspects.
- Bridge to theory, while answering pragmatic concerns: We need to bridge properties/qualities of interpretable models and of explanations provided with fundamental theories and facts about human cognition and the qualities of explanations, so as to understand better the mechanisms for providing context-dependent qualitative explanations.
- Focus on transparency: There is not any comprehensive XDRL framework which has been developed for providing explanations under pragmatic constraints. Although there are approaches targeting to transparency, the term is used more or less equivalently to interpretability. We need methods/frameworks whose explainability capabilities can be tuned to pragmatic constraints, focusing on transparency.
- Explore the interpretable box design paradigm: We need to further explore the distinct XDRL paradigms for interpretable box design, also in conjunction to other objectives (e.g. as in (VIPER, )): While the mimicking paradigm provides a solution that somehow by-passes explaining the deep models in a direct way, it adds a layer to the system that is susceptible to additional loss of accuracy in problem solving and raises questions about its fidelity to the control logic of the DRL agent. Models’ distillation provides opportunities for generalizing effectively beyond the training examples provided originally to the agent and allow building compact models. However, we need to explore further methods for distilling acquired knowledge in direct ways, and exploiting these models to provide concise explanations’ for different purposes and across problem tasks.
- Develop XDRL in a principled way: We need principles, methodologies and tools for designing and developing XDRL agents from early development phases. While there are methods that follow the interpretable box design paradigm, which is close to that aim, most approaches add explainability layers/modules to agents with advanced DRL methods.
- Explore XDRL in complex and/or large-scale multi-agent settings: Solving explainability problems for multi-agent (MAS) and potentially complex settings, does not follow in a straightforward way from solving these problems for individual agents: Interpretability may involve further filtering/aggregation steps in providing explanations’ content, while we need to advance explainability and transparency in challenging but highly critical real-world MAS settings.
- Set up experimental protocols with emphasis on pragmatic concerns: We need setting up experimental protocols, methodologies and widely acceptable indicators regarding how and what one should evaluate to provide evidence regarding humans’ qualitative understanding and acceptability of explanations, closely connected to the functionality of DRL agents deployed in real-world settings. A.Holzinger at al. in (ExplanationsQuality, ) propose System Causability Scale for measuring explanations’ quality, that can be a starting point, incorporating issues of transparency.
- Provide comparative evaluations for interpretability, explainability and transparency: Although some kinds of models are considered more interpretable than others, or some explainability methods may provide better explanations than others (whatever this means, but w.r.t the explanation problem addressed) we need a comprehensive comparative evaluation on the quality of explanations that interpretable models can provide, also w.r.t pragmatic issues.
Acknowledgements.
This work has been partially supported by the TAPAS H2020-SESAR-2019-2 Project (GA number 892358) Towards an Automated and exPlainable Air traffic management (ATM) System.References
- (1) Puiutta, E. and E. MSP Veith, “Explainable Reinforcement Learning: A Survey”. arXiv:2005.06247, 2020.
- (2) Ribeiro, M. T., S. Singh, and C. Guestrin. “Why Should I Trust You?: Explaining the Predictions of Any Classifier”. In Proceedings of the 22nd ACM SIGKDD KDD. ACM, New York, NY, USA, 1135–1144. 2016.
- (3) Iyer, R., Y. Li, H. Li, M. Lewis, R. Sundar and K. Sycara. “Transparency and Explanation in Deep Reinforcement Learning Neural Networks”. AAAI/ACM Conference on AI, Ethics, and Society 2018, also in arXiv:1809.06061, 2018.
- (4) Pocius, R., L. Neal, A. Fern, “Strategic Tasks for Explainable Reinforcement Learning”. Vol 33 No 01: AAAI-19, IAAI-19, EAAI-20, 2019.
- (5) Shi W., Song S., Wang Z. and Huang G. “Self-Supervised Discovering of Causal Features: Towards interpretable Reinforcement Learning”. arXiv:2003.07069v2, 2020.
- (6) Juozapaitis, Z., A. Koul, A. Fern, M. Erwig, and F. Doshi-Velez. “Explainable reinforcement learning via reward decomposition”. In IJCAI/ECAI Workshop on Explainable Artificial Intelligence, 2019.
- (7) Topin, N., M. Veloso. “Generation of Policy-Level Explanations for Reinforcement Learning”. AAAI, arXiv:1905.12044. 2019.
- (8) Goldberger, J., G. E. Hinton, S. T. Roweis, and R. R. Salakhutdinov. “Neighbourhood components analysis.” NIPS, pp. 513-520. 2005.
- (9) Cruz, F, R. Dazeley, and P. Vamplew. “Memory-Based Explainable Reinforcement Learning.” In AJCAI, pp. 66-77. 2019.
- (10) Leech, T. “Explainable machine learning for task planning in robotics.” MSc Thesis, MIT, 2019.
- (11) Dancey, D., Z. A. Bandar, and D. McLean. ”Logistic model tree extraction from artificial neural networks.” IEEE Trans. on Syst., Man, and Cybern., Part B, Vol. 37, no. 4, 794-802. 2007.
- (12) Liu, G., O. Schulte, W. Zhu, and Q. Li. ”Toward interpretable deep reinforcement learning with linear model u-trees.” ECML PKDD. 2018.
- (13) Che, Z., S. Purushotham, R. Khemani, and Y. Liu. ”Interpretable deep models for ICU outcome prediction.” In Proc. of AMIA Annual Symp. 2016.
- (14) Wang, N., D. V. Pynadath, and S. G. Hill. ”The Impact of POMDP-Generated Explanations on Trust and Performance in Human-Robot Teams.” AAMAS’16. 2016.
- (15) D. V. Pynadath and S. C. Marsella. “Fitting and Compilation of Multiagent Models through Piecewise Linear Functions”. AAMAS ’04. 2004.
- (16) Madumal, P., T. Miller, L. Sonenberg, and F. Vetere. ”Explainable reinforcement learning through a causal lens.” arXiv:1905.10958. 2019.
- (17) J. Y. Halpern and J. Pearl. “Causes and explanations: A structural-model approach. Part I: Causes.” BJPS, 2020.
- (18) Coppens, Y., K. Efthymiadis, T. Lenaerts, A. Nowé, T. Miller, R. Weber, and D. Magazzeni. ”Distilling deep reinforcement learning policies in soft decision trees.” IJCAI 2019 Workshop on Explainable Artificial Intelligence. 2019.
- (19) Frosst, N. and G. Hinton. “Distilling a Neural Network Into a Soft Decision Tree”. In Proc. of the 1st Inter. Workshop on Comprehensibility and Explanation in AI and ML, vol. 2071 of AI*IA Series at CEUR Workshop Proceedings. 2017.
- (20) Madumal, P., T. Miller, L. Sonenberg, and F. Vetere. ”Distal Explanations for Explainable RL Agents.” arXiv:2001.10284. 2020.
- (21) Roth, A. M., N. Topin, P. Jamshidi, and M. Veloso. ”Conservative Q-Improvement: Reinforcement Learning for an interpretable Decision-Tree Policy.” arXiv:1907.01180. 2019.
- (22) Fukuchi, Y., M. Osawa, H. Yamakawa, and M. Imai. ”Autonomous self-explanation of behavior for interactive reinforcement learning agents.” In Proc. of the 5th Intl. Conf. on Human Agent Interaction. 2017.
- (23) Hayes, B., and J. A. Shah. ”Improving robot controller transparency through autonomous policy explanation.” In 12th ACM/IEEE Intl. Conf. on Human-Robot Interaction. 2017.
- (24) Mohseni, S., N. Zarei, and E. D. Ragan. ”A Multidisciplinary Survey and Framework for Design and Evaluation of Explainable AI Systems.” arXiv-1811.11839. 2018.
- (25) Guidotti, R., A. Monreale, S. Ruggieri, F. Turini, F. Giannotti, and D. Pedreschi. ”A survey of methods for explaining black box models.” ACM computing surveys (CSUR) 51, no. 5 (2018): 1-42. 2018.
- (26) Lundberg, S. M., and S. Lee. ”A unified approach to interpreting model predictions.” NIPS’17. 2017.
- (27) Verma, A., V. Murali, R. Singh, P. Kohli, and S. Chaudhuri. ”Programmatically interpretable reinforcement learning.” arXiv:1804.02477. 2018.
- (28) Shu, T., C. Xiong, and R. Socher. ”Hierarchical and interpretable skill acquisition in multi-task reinforcement learning.” ICLR 2018.
- (29) Boz, O. ”Extracting decision trees from trained neural networks.” In Proc. of ACM SIGKDD KDD’02, pp. 456-461. 2002
- (30) Huang, S. H., D. Held, P. Abbeel, and A. D. Dragan. ”Enabling robots to communicate their objectives.” Autonomous Robots 43, no. 2, 309-326. 2019.
- (31) Ribeiro, M. T., S. Singh, and C. Guestrin. “Anchors: High-precision model-agnostic explanations”. In Proc. of AAAI’18. 2018.
- (32) Guidotti, R., A. Monreale, S. Ruggieri, D. Pedreschi, F. Turini and F. Giannotti. “Local rule-based explanations of black box decision systems”. arXiv:1805.10820. 2018
- (33) Shrikumar, A., et al. “Not Just a Black Box: Learning Important Features Through Propagating Activation Differences”. arXiv:1605.01713. 2016
- (34) Bach, S., et al. “On pixel-wise explanations for non-linear classifier decisions by layerwise relevance propagation”. PloS One 10.7. 2015.
- (35) Pyeatt, L. D., and A. E. Howe. “Decision tree function approximation in reinforcement learning.” 3rd international symposium on adaptive systems: evolutionary computation and probabilistic graphical models. Vol. 2. No. 1/2. 2001
- (36) McCallum, A.K., et al. “Learning to use selective attention and short-term memory in sequential tasks”. In Proc. SAB. vol. 4, pp. 315-325. 1996
- (37) Uther, W.T., M.M. Veloso. “Tree based discretization for continuous state space reinforcement learning”. AAAI/IAAI. pp. 769-774. 1998
- (38) Rizzo, S. G., G. Vantini and S. Chawla. “Reinforcement Learning with Explainability for Traffic Signal Control,” IEEE ITSC, pp. 3567-3572, 2019.
- (39) Lipton, Z. C. “The Mythos of Model Interpretability: In machine learning, the concept of interpretability is both important and slippery”. Queue 16, 3 (May-June 2018), 31–57. 2018.
- (40) Biran, O., and C. V. Cotton. “Explanation and Justification in Machine Learning : A Survey”. IJCAI-17 workshop on explainable AI (XAI). 2017
- (41) de Graaf, M.A., B. F. Malle, A. Dragan, and T. Ziemke. “Explainable Robotic Systems”. In Companion of ACM/IEEE HRI ’18, 387–388. 2018.
- (42) Miller, T. “Explanation in artificial intelligence: Insights from the social sciences”. Artificial Intelligence, Vol. 267, pp 1-38. 2019.
- (43) Amir, D., and O. Amir. “HIGHLIGHTS: Summarizing Agent Behavior to People”. AAMAS ’18. 2018.
- (44) Sequeira, P., and M. Gervasio. “Interestingness Elements for Explainable Reinforcement Learning: Understanding Agents’ Capabilities and Limitations.” Artificial Intelligence Vol. 288. Also in arXiv:1912.09007v2. 2020.
- (45) Rusu, A. A., et. al. “Policy Distillation”. arXiv:1511.06295v2. 2015.
- (46) Hinton, G., O. Vinyals, and J. Dean. “Distilling the Knowledge in a Neural Network.” Deep Learning and Representation Learning Works., NIPS, 2014.
- (47) Huang, S. H., K. Bhatia. P. Abbeel and A. D. Dragan. ”Establishing Appropriate Trust via Critical States”. arXiv:1810.08174, 2018.
- (48) van der Waa, J., J. van Diggelen, K. van den Bosch and M. Neerincx. “Contrastive Explanations for Reinforcement Learning in terms of Expected Consequences”. arXiv:1807.08706. 2018.
- (49) National Security Commission on Artificial Intelligence. Interim Report. November 2019.
- (50) EU High-level expert group on Artificial Intelligence. Ethics Guidelines for Trustworthy Artificial Intelligence. April 8, 2019. Accessible in file:///Users/VG/Downloads/AIHLEG_EthicsGuidelinesforTrustworthyAI-ENpdf.pdf
- (51) Annasamy, R. M., and K. Sycara. ”Towards better interpretability in deep q-networks.” AAAI’19. Vol. 33. No. 01. 2019.
- (52) Mott, A., et al. ”Towards interpretable reinforcement learning using attention augmented agents.” arXiv:1906.02500. 2019.
- (53) Zahavy, T., N. Ben Zrihem and S. Mannor. “Graying the black box: Understanding DQNs.” arXiv:1602.02658. 2017
- (54) Bastani, O., Y. Pu and A. Solar-Lezama. “Verifiable Reinforcement Learning via Policy Extraction.” arXiv:1805.08328. 2019
- (55) Murdoch, W. J., C. Singh, K. Kumbier, R. Abbasi-Asl and B. Yu. “Definitions, methods, and applications in interpretable machine learning”, Proc. of the National Academy of Sciences 116(44), 22071–22080. 2019.
- (56) Holzinger, A., A. Carrington, and H. Meuller. “Measuring the Quality of Explanations: The System Causability Scale (SCS)”. KI - Künstliche Intelligenz 34, 193–198. 2020.
- (57) Rudin, C., C. Chen, Z. Chen, H. Huang, L. Semenova and C. Zhong. “Interpretable Machine Learning: Fundamental Principles and 10 Grand Challenges”. arXiv:2103.11251v2. 2021.
- (58) Belle, V. and I. Papantonis. “Principles and Practice of Explainable Machine Learning”. arxiv: 2009.11698v1. 2020
- (59) Yau, H., C. Russell and S. Hadfield. “What Did You Think Would Happen? Explaining Agent Behaviour Through Intended Outcomes”. NeurIPS 2020.
- (60) Atrey, A., K. Clary, and D. Jensen. “Exploratory not explanatory: Counterfactual analysis of saliency maps for deep reinforcement learning”. arXiv:1912.05743. 2019
- (61) Hüyük, A., D. Jarrett, C. Tekin, and M. van der Schaar. “Explaining by Imitating: Understanding Decisions by Interpretable Policy Learning”. ICLR, 2020.
- (62) Gottesman, Omer and Futoma, Joseph and Liu, Yao and Parbhoo, Sonali and Celi, Leo and Brunskill, Emma and Doshi-Velez, Finale. “Interpretable off-policy evaluation in reinforcement learning by highlighting influential transitions”. ICML, PMLR 2020
- (63) Puri, N., S. Verma, P. Gupta, D. Kayastha, S. Deshmukh, B. Krishnamurthy and S. Singh. “Explain Your Move: Understanding Agent Actions Using Specific and Relevant Feature Attribution”. ICLR 2020, also in arXiv:1912.12191. 2020.
- (64) Huber, T., K. Weitz, E. André and O. Amir. “Local and Global Explanations of Agent Behavior: Integrating Strategy Summaries with Saliency Maps”. arXiv:2005.08874. 2021
- (65) Skirzyński, J., F. Becker, and F. Lieder. “Automatic discovery of interpretable planning strategies”. Machine Learning. 2021.
- (66) Greydanus, S., A. Koul, J. Dodge, and A. Fern. “Visualizing and Understanding Atari Agents”. ICML, PLMR 80, 2018
- (67) Silver, T., R. A. Kelsey, A. K. Lew, L. P. Kaelbling and J. Tenenbaum. “Few-Shot Bayesian Imitation Learning with Logical Program Policies”. AAAI 2020 and arxiv 1904.06317. 2019
- (68) Alqaraawi, A., M. Schuessler, P. Weiss, E. Costanza and N. Berthouze. “Evaluating saliency map explanations for convolutional neural networks: A user study”. arXiv:2002.00772. 2020
- (69) Erwig, M., A. Fern, M. Murali and A. Koul. ”Explaining Deep Adaptive Programs via Reward Decomposition”. IJCAI/ECAI Workshop on XAI. 2018
- (70) Huber, T., D. Schiller and E. André. “Enhancing Explainability of Deep Reinforcement Learning Through Selective Layer-Wise Relevance Propagation”. In Proc. of KI 2019: Advances in Artificial Intelligence. 2019.
Appendix A XDRL state of the art: Methods characteristics and evaluation
This supplementary part describes in a succinct and comprehensive way the XDRL methods reviewed: The aim is to identify the main characteristics of each method, as these have been described in Section 4, specifying the explainability problem each method aims to solve, the XDRL paradigm it follows, what is the explanation content provided, and how this is surfaced, according to the proposal made. In conjunction to that information, and to further help identifying the most appropriate method to be applied in a specific context, this section provides XDRL requirements for the DRL methods used, as well as “XDRL - DRL interplay aspects”: Access to DRL models’ information and/or samples required, as well as whether interpretable XDRL models need to access this information in an online mode (i.e., while the DRL method is being trained, or while the DRL agent takes decisions) or in an offline mode (i.e., taking batches of required samples and data from DRL models).
Specifically, three tables summarize the XDRL approaches: Table LABEL:tab:overall provides an overview of the approaches reviewed, starting from methods addressing the model inspection problem, proceeding to those approaches that address the global policy and objectives’ model explanation problems, and then, to those addressing the local response/objectives’ explanation problem. Groups of methods are separated by a bold line. Table LABEL:tab:overall identifies per method the paradigm followed, the interpretable model constructed, the required access to the DRL models via the function (i.e., the input to the interpretation process), as well as the output of the interpretation process - the explanation content, and finally the surface representation of explanations. Table LABEL:tab:evaluation provides an overview of evaluating each of the methods, focusing on evaluation goals and assessing the quality of explanations. It particularly specifies the involvement of humans in the evaluation process, while it provides information on the way explanations are evaluated and the objectives of that evaluation. It repeats the content of explanations provided per method from Table LABEL:tab:overall, in order to be made self-contained. Finally, Table LABEL:tab:functionality provides the profile of the DRL methods to which each XDRL approach is compatible with or with which it has been demonstrated/evaluated with. It particularly identifies the DRL method used in the article proposing the XDRL method (if any, else it identifies again the specific XDRL model); the DRL input assumed, as some of the approaches are assuming visual only input; the state and action space requirements; and finally, whether XDRL is able to provide the explanation content in an online or in an offline manner, according to the distinction made above.
Important detailed remarks that can be made following the tabular presentation of methods, are as follows:
According to the mimicking/distillation distinction we made, the approaches that follow the distillation paradigm are underrepresented: Thus, we need more approaches that exploit DRL models (of whatever kind) in a direct way for providing explanation content and answering various questions in high fidelity, and at different levels of scale and granularity, addressing all explainability problems.
Following this remark, attention models provide a promising approach towards this direction, but approaches using such models follow the interpretable box design paradigm, only. More importantly, attention models have been used mostly for visual agents (this holds for approaches that aim at solving the policy explanation and the outcome explanation problems).
Abstractions of states, transitions and trajectories offer promising generic approaches towards providing explanations at various levels of granularity and scale, and their applicability should be extended to arbitrary hierarchical levels of abstraction, exploiting also the possibilities of DRL methods to learn hierarchical models.
Approaches that offer examples of agent’s behaviour in appropriate contexts and under specific constraints, or that highlight important/critical aspects affecting agent’s behaviour, have been proposed towards building trust to agents’ capabilities, or for training the humans on agents’ capabilities, preferences and limitations. Such interesting approaches have been reported for the policy explanation, objective explanation and the outcome explanation problems.
Finally, regarding evaluation of explanations, there are numerous objectives and ways to measure effectiveness that deem the approaches practically incomparable: This is a side affect of the lack of agreed definitions on terms and problems to be solved as far as explainability, interpretability and transparency are concerned. Further on that, as far as transparency is concerned, there is lack of an approach that aims at addressing pragmatic concerns to the deployment of XDRL agents, while the involvement of humans in experiments in any of the reported works does not concern the involvement of operators in (simulated) real world settings.
Method | Problem addressed | Paradigm | Interpretable Model | Model aspects exploited by | Explanation | Explanation |
[Ref](Sect.) | ( - model) | (arguments) | content | surface representation | ||
(LIME, )(4.1.1) | Model inspection (LIME) and model explanation (SP-LIME) | Generic | linear model, decision trees, etc. | Problem instances’ features | (a) Important features (b) Examples of cases | Visual |
(SHAP, )(4.1.2) | Model inspection and response explanation | Generic | linear model | Problem instances’ features | Important features | Visual |
(Highlights, )(4.2.1) | Policy explanation | Generic | Highlights (sub-trajectories with important states) | Trajectories, Q values | Sub-trajectories exemplifying agent behaviour | Visual |
(LocalAndGlobal, )(4.2.1) | Policy and outcome explanation | Generic | Highlights (policy) and LRP (Outcome) | Trajectories, Q values, NN weights | Sub-trajectories and saliency of features | Visual |
(ContrastiveExplanations, )(4.2.2) | Policy explanation problem | Generic | Qualitative descriptions of policies rollouts | MDP, Q values, rewards | Trajectories in terms of user-interpretable descriptions of states and actions | Natural language (template-based) |
(CriticalStates, )(4.2.3) | Policy explanation | Generic | Critical states and corresponding actions | Q values | Critical states and actions | Visual |
(GrayingBlackBox, )(4.2.4) | Policy explanation and inspection | Generic | Approximating the MDP via SAMDP | State Features, rewards, neural activations of states | State clusters’ and SAMDP | Visual inspection of states’ clusters and SAMDP via t-SNE |
(betterInterpretability, )(4.2.5) | Policy explanation | Interpretable box design | Attention model | None | State representation clusters’ and reconstructed images for specific key-value pairs | Visual (state embeddings’ clusters via t-SNE, and reconstructed images) |
(InterpretableOPE, )(4.2.6) | Policy explanation | Generic | Influence Functions | None | Influences of transitions | Visual representation of trajectories and influences of transitions |
(AttAugmentedAgents, )(4.2.7) | Policy explanation | Interpretable box design | Attention model | None | Input frames with attention maps | Visual (attention maps, saliency maps) |
(ConservativeQImprovement, )(4.2.8) | Policy explanation | Interpretable box design | Decision Tree | None | Decision Trees’ constructs | Not specified |
(HierarchicalinterpretableRL, )(4.2.9) | Policy explanation | Interpretable box design | Hierarchical plans with natural language descriptions of skills/tasks and Stochastic temporal grammar (STG) | None | Hierarchical plans with language-grounded tasks | Natural language |
(LVIN, )(4.2.10) | Model explanation | Mimicking | High-level FSA and low-level MDP | Raw model (deep NN) weights | Transition matrices / graphs, and generated trajectories | Not specified |
(interpretableDRLwithLMUTs, )(4.2.11) | Model explanation | Mimicking | Linear Model U-Trees (LMUTs) | DRL decisions, and Q values | Decision making rules using partition cells and saliency maps | Not specified |
(PolicyLevelExplanations, )(4.2.12) | Policy explanation | Mimicking | Abstracted Policy Graphs (Markov chains of abstract states) | Abstracted Policy Graphs and Features’ importance | Conditional expected scores of states, Value function | Not specified |
(PIRL, )(4.2.13) | Policy explanation | Mimicking | Programmatic policies | None | Functional program | Functional program |
(ExplainingByImitation, )(4.2.14) | Policy explanation | Mimicking | Decision dynamics and decision boundaries | None | Action-belief trajectories | Visualization of trajectories in belief complexes |
(InterpretablePlanningStrategies, )(4.2.15) | Policy explanation | Mimicking | Logical Program Policies (LPP) | None | Logical rules in disjunctive normal form | Flawcharts representing the decision rules |
(CommunicateObjectives, )(4.3.1) | Objectives explanation | Generic | Linear objective function parameters; Models of human beliefs on parameters | Examples of trajectories | Trajectories | Not specified |
(WhatDidYouThink, )(4.3.2) | Objectives explanation | Generic | Belief maps (discounted visitations of future state-action pairs) | Discounted visitations and Q-values | Belief maps | Visualizations of belief maps (hit maps) |
(DistalExplanations, )(4.3.3) | Policy and Objectives explanation | Mimicking | Decision tree (policy) and causal model (objectives), opportunity chains (importance of responses) | None | Decision tree nodes, causal chain variables and distal action | Natural language |
(RewardDecomposition, )(4.4.1) | Outcome explanation | Generic | Difference explanations | Q values | Decomposed Q values with reward difference and minimal sufficient explanations | Visual |
(AutonomousPolicyExplanation, )(4.4.2) | Outcome explanation | Generic | Control logic (states by means of variables and actions by annotated functions | States’ features | Representation of states’ features | Natural language (template-based) using communicable predicates |
(TransparencyCommunicationforRL, )(4.4.3) | Outcome explanation | Generic | POMDP components | Agent’s beliefs, transition models, Q values | POMDP components (updates) | Natural language (template-based) |
(TransparencyandExplanationInDRL, )(4.4.4) | Outcome explanation | Generic | Pixel/Object salience | Q values | Recognized objects’ salience | Visual |
(ExplainYourMove, )(4.4.5) | Outcome explanation | Generic | Features’ salience | Q values | Features salience (specificity, relevance) | Visual |
(AutonomousSelfExplanation, )(4.4.6) | Outcome explanation | Generic | Target states and expert instructions | Experts’ instructions | Mappings of changes to instructions | Visual |
(CausalLens, )(4.4.7) | Response and Objectives explanation | Mimicking | Structural equations (quantitative effects of actions to variables (approximated by multivariate regression models) | None | Causal chains in action influence models exploiting structural equations | Not specified |
(Self-SupervisedDiscoveryOfCausalFeatures, )(4.4.8) | Outcome explanation | Mimicking | Attention masks highlighting important feature states | Policy-generated state-action pairs | Attention masks | Visual (applies to vision-based RL) |
(DistillingDRLInSDTs, )(4.4.9) | Outcome explanation | Mimicking | Soft Decision Tree | None | Learned filters along the traverse path in SDT | Visual (hitmaps) |
(MemoryBAsedXRL, )(4.4.10) | Outcome explanation | Distillation | Past transitions | Q values | Probabilities of choosing actions and probabilities of success | Visual |
(LRP-argmax, )(4.4.11) | Outcome explanation | Generic | NN units relevance values | NN weights | Input relevance (saliency) | Saliency maps |
Method | Involvement | Content of explanations | Explanations’ quality | Evaluation Objectives |
[Ref](Sec.) | of humans | measurement | ||
(LIME, )(4.1.1) | Yes | (a) Important features (b) Examples of cases | Faithfulness and Trustworthiness given a set of gold features to be recovered | Increasing trustworthiness, improving classifiers |
(SHAP, )(4.1.2) | Yes | Important features | Consistency with human intuition, ability to explain subtle classification differences | Sample efficiency; Understanding classifiers’ abilities |
(Highlights, )(4.2.1) | Yes | Sub-trajectories exemplifying agents’ behaviour | Helpful in assessing agents’ capabilities and trustworthiness | Increasing familiarity with agents’ capabilities and limitations |
(LocalAndGlobal, )(4.2.1) | Yes | Sub-trajectories and saliency maps | Modeling agents’ behaviour, capabilities, attention and trustworthiness | Building models of agents’ behaviour, trusting agents and assessing the role of saliency maps in policy explanations |
(ContrastiveExplanations, )(4.2.2) | Yes | Trajectories in terms of user-interpretable descriptions of states and actions | Properties of explanations (length, amount of information, time-horizon of consequences) | Provide contrastive explanations in a comprehensive way |
(CriticalStates, )(4.2.3) | Yes | Critical states and actions | Subjective and objective development of humans’ trust | Establishing trust on agents’ limitations and capabilities |
(GrayingBlackBox, )(4.2.4) | No | State clusters’ and SAMDP visualizations | Effectiveness in analysing DQN’s policy at different levels of abstraction | Analysing DQN’s policy |
(betterInterpretability, )(4.2.5) | No | State representation clusters’ and reconstructed images for specific key-value pairs | Effectiveness in task execution comparable to deep Q-learning methods | Interpretable visual Q-learning |
(InterpretableOPE, )(4.2.6) | Yes | Influence of transitions | Influence analysis | Evaluating the trustworthiness of OPE methods for noisy and confounded observational data |
(AttAugmentedAgents, )(4.2.7) | No | Input frames with attention maps | Predictability, comprehensibility and focus of attention | Interpretable visual DRL |
(ConservativeQImprovement, )(4.2.8) | No | Decision Trees’ constructs | Quality of the policy (modelled by a decision tree) an model size | Interpretable Q-learning |
(HierarchicalinterpretableRL, )(4.2.9) | No | Hierarchical plans with language-grounded tasks | None | Policy learning, generalization and interpretability |
(LVIN, )(4.2.10) | No | Transition matrices / graphs, and generated trajectories | Not measured | Imitation; Interpretability, verifiability and manipulability of the model |
(interpretableDRLwithLMUTs, )(4.2.11) | No | Decision making rules using partition cells and saliency maps | Fidelity to the DQN model decisions, samping efficiency and effectiveness in performing tasks | Interpretability and direct task performance |
(PolicyLevelExplanations, )(4.2.12) | No | Abstracted Policy Graphs (Markov chains of abstract states) and Features’ importance | Prediction of features’ importance and explanation (model) size | Concise policy-level explanations’ model in tractable time |
(PIRL, )(4.2.13) | No | Functional program | Performant and short programmatic policies | Interpretability and verfication of programmatic policies |
(ExplainingByImitation, )(4.2.14) | Yes | Action-belief trajectories | Interpretability, Accuracy, Subjectivity | Understanding decisions dynamics |
(InterpretablePlanningStrategies, )(4.2.15) | Yes | Logical Program Policies (Logical rules in disjunctive normal form) | Expected score of the plan and agreement withe conveyed policy | Improving human planning quality and decision making by conveying policies as flowcharts |
(CommunicateObjectives, )(4.3.1) | Yes | Examples of trajectories | None | Understanding agents’ behaviour via modeling the objective function |
(WhatDidYouThink, )(4.3.2) | No | Belief maps (discounted valuations of future state-action pairs) | Not measured | Consistency of intended outcome with Q values |
(DistalExplanations, )(4.3.3) | Yes | Decision tree nodes, causal chain variables and distal action | Subjective evaluation of explanation quality and task prediction | Understanding of the agent and provision of better explanations |
(RewardDecomposition, )(4.4.1) | Yes | Decomposed Q values with reward difference and minimal sufficient explanations | None | Understanding agent decisions for delving into DRL issues |
(AutonomousPolicyExplanation, )(4.4.2) | No | Representation of states’ features | No measures: Subjective comparison with experts’ explanations | Interpretability of control logic |
(TransparencyCommunicationforRL, )(4.4.3) | No | POMDP components | None | Interpretability and repairing trust |
(TransparencyandExplanationInDRL, )(4.4.4) | Yes | Recognized objects’ salience | Correctness of matching and predictions agents’ behaviour; consistency of explanations with saliency maps | Explaining the influence of objects in agents’ decisions |
(ExplainYourMove, )(4.4.5) | Yes | State features’ salience | Salience (qualitative), identifying specific and relevant features (accuracy), robustness to pertrubations | Understanding agent actions |
(AutonomousSelfExplanation, )(4.4.6) | Yes | Mappings of changes to instructions | None | Predictability of agents’ behaviour |
(CausalLens, )(4.4.7) | Yes | Causal chains in action influence models exploiting structural equations | Subjective evaluation for aggregate video explanations of explanation models | Better understanding of agent strategies and promoting trust |
(Self-SupervisedDiscoveryOfCausalFeatures, )(4.4.8) | No | Attention masks | Feature Overlapping Rate and Background Elimination Rate | Enhanced explanation of visual agents’ decision making process |
(DistillingDRLInSDTs, )(4.4.9) | No | Learned filters along the traverse path in SDT | None | Understanding the policy w.r.t to state features |
(MemoryBAsedXRL, )(4.4.10) | No | Probabilities of choosing actions and probabilities of success | None | Explaining decision making w.r.t to the intended task |
(LRP-argmax, )(4.4.11) | No | Relevance values of inputs | None | Selective and inclusive saliency maps |
Method | DRL model | DRL Input | State / Action Space | XDRL Learning Mode w.r.t DRL |
[Ref](Sec.) | (Visual, etc) | (Discrete / Continue) | (Online, Incremental or Offline) | |
(LIME, )(4.1.1) | Classification/prediction model | Generic | Interpretable representations of state features | Offline |
(SHAP, )(4.1.2) | Classification/prediction model | Generic | Interpretable representations of state features | Offline |
(Highlights, )(4.2.1) | Policy model | Generic | Discrete actions and states | Online |
(LocalAndGlobal, )(4.2.1) | Policy model | Visual | Discrete actions | Offline |
(ContrastiveExplanations, )(4.2.2) | Policy model | Generic | Discrete actions | Online |
(CriticalStates, )(4.2.3) | No Model | Generic ( although visual seems more appropriate) | Discrete actions and state features | Online |
(GrayingBlackBox, )(4.2.4) | Policy model in Q-Learning | Generic | Discrete actions | Exploration |
(betterInterpretability, )(4.2.5) | Policy model in Q-Learning (DQN) | Visual | Discrete actions but discrete/continuous state features | Directed exploration |
(InterpretableOPE, )(4.2.6) | Policy model | Generic | Generic | Offline |
(AttAugmentedAgents, )(4.2.7) | Policy model (LSTM) | Visual | Discrete actions | Exploration |
(ConservativeQImprovement, )(4.2.8) | Policy model (Decision tree) | Generic | Discrete actions and states | Exploration |
(HierarchicalinterpretableRL, )(4.2.9) | Hierarchical policy model | Visual | Discrete actions | Curriculum learning |
(LVIN, )(4.2.10) | Policy model (CNN) | Generic (restricted to 2D space) | Discrete actions | Offline |
(interpretableDRLwithLMUTs, )(4.2.11) | Deep policy model in Q-Learning (DQN) | Generic | Discrete actions | Online (active play) and Offline (batch mode) |
(PolicyLevelExplanations, )(4.2.12) | Policy model | Generic | Discrete actions and binary state features | Offline |
(PIRL, )(4.2.13) | Policy model (DDPG) | Generic | Discrete or continuous actions and states | Offline in periodic batches of histories via input augmentation |
(ExplainingByImitation, )(4.2.14) | No Model | Generic | Generic | Offline |
(InterpretablePlanningStrategies, )(4.2.15) | No Model | Generic | Generic | Offline |
(CommunicateObjectives, )(4.3.1) | Linear objective function | Generic | Generic | Offline |
(WhatDidYouThink, )(4.3.2) | Value-based (Q-learning) | Generic (low-dimensional space with deterministic MDP) | Discrete state-action space | Online and Offline |
(DistalExplanations, )(4.3.3) | Policy (Decision trees), distal action model and objectives (causal model) | Generic | Discrete Actions and states | Online |
(RewardDecomposition, )(4.4.1) | Decomposed reward and policy (drDQN, drSARSA) | Generic | Discrete Actions | Online |
(AutonomousPolicyExplanation, )(4.4.2) | Policy and domain model | Generic | Discrete actions and discrete or continuous states | Offline |
(TransparencyCommunicationforRL, )(4.4.3) | POMDP components (focusing on the Observations model) | Generic | Discrete actions and states | Offline |
(TransparencyandExplanationInDRL, )(4.4.4) | Policy model | Visual | Generic | Online |
(ExplainYourMove, )(4.4.5) | Q values model | Visual | Discrete states and actions | Online |
(AutonomousSelfExplanation, )(4.4.6) | Policy model | Generic | Generic | Online |
(CausalLens, )(4.4.7) | Policy model | Generic | Discrete actions | Online |
(Self-SupervisedDiscoveryOfCausalFeatures, )(4.4.8) | Policy model (PPO, SAC, TD3) | Visual | Generic | Offline |
(DistillingDRLInSDTs, )(4.4.9) | Policy model (PPO) | Spatial | Generic | Offline |
(MemoryBAsedXRL, )(4.4.10) | Policy model (SARSA) | Generic | Discrete actions and states | Offline |
(LRP-argmax, )(4.4.11) | Policy model | Generic (applied to visual) | Discrete actions | Online |