Next Article in Journal
Dynamic Path Planning Based on 3D Cloud Recognition for an Assistive Bathing Robot
Previous Article in Journal
FI-NPI: Exploring Optimal Control in Parallel Platform Systems
Previous Article in Special Issue
Architectural Proposal for Low-Cost Brain–Computer Interfaces with ROS Systems for the Control of Robotic Arms in Autonomous Wheelchairs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Causal-Based Approaches to Explain and Learn from Self-Extension—A Review

1
Departamento Tecnología Electrónica, University of Málaga, 29071 Málaga, Spain
2
Departamento Tecnologías de los Computadores y las Comunicaciones, University of Extremadura, 10005 Cáceres, Spain
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(7), 1169; https://doi.org/10.3390/electronics13071169
Submission received: 16 January 2024 / Revised: 8 March 2024 / Accepted: 17 March 2024 / Published: 22 March 2024
(This article belongs to the Special Issue Intelligent Control and Computing in Advanced Robotics)

Abstract

:
The last decades have seen a revolution in autonomous robotics. Deep learning approaches and their hardware implementations have made it possible to endow robots with extraordinary perceptual capabilities. In addition, they can benefit from advances in Automated Planning, allowing them to autonomously solve complex tasks. However, on many occasions, the robot still acts without internalising and understanding the reasons behind a perception or an action, beyond an immediate response to a current state of the context. This gap results in limitations that affect its performance, reliability, and trustworthiness. Deep learning alone cannot bridge this gap because the reasons behind behaviour, when it emanates from a model in which the world is a black-box, are not accessible. What is really needed is an underlying architecture based on deeper reasoning. Among other issues, this architecture should enable the robot to generate explanations, allowing people to know why the robot is performing, or has performed, a certain action, or the reasons that may have caused a certain plan failure or perceptual anomaly. Furthermore, when these explanations arise from a cognitive process and are shared, and thus validated, with people, the robot should be able to incorporate these explanations into its knowledge base, and thus use this understanding to improve future behaviour. Our article looks at recent advances in the development of self-aware, self-evolving robots. These robots are designed to provide the necessary explanations to their human counterparts, thereby enhancing their functional capabilities in the quest to gain their trust.

1. Introduction

The upcoming Society 5.0 represents a new paradigm in which people and artificial beings cooperate in daily life routines, environments, and interactions [1]. This cooperation is intended to be natural and intuitive, and the new artificial actors are expected to behave appropriately. However, the concept of what is an ’appropriate’ behaviour is context-specific and just as complex and diverse as societies and individuals are. Moreover, it depends not only on functionality but also on acceptability, accessibility, and adaptability [2,3]. Service robots constitute one of the technologies with the highest number of potential applications and are also one of the most deeply affected by these technical and contextual complexities [4]. Service robots, being inevitably social when used in daily life scenarios, face the most challenging technical, ethical, social, and legislative demands [5].
Socially-aware Robotics (SAWR) is an emergent area of research that seeks to understand how cognitive robots can become aware of their social context and use this capability to behave as more accessible, accepted, and useful devices, being able to establish more adequate and effective interactions to help humans [6]. A robot aiming to exploit autonomous capabilities that are socially enhanced needs to sense its environment and achieve a certain level of understanding about what happens within this environment. However, to be truly socially aware, a robot must not only purposely react to perceived changes, it should also predict the incoming events and anticipate their consequences, selecting the best possible and most understandable action and respecting social conventions [7,8]. Their behaviour must also be correctly interpreted by the people with whom they share their environment. The process of anticipation reduces the response time and has the potential to be perceived by humans as more natural, increasing the chances for the robot to produce a socially adequate type of behaviour. These robot–environment dynamics anchor the robot to its world and open the way to online autonomous learning. Thus, socially aware robots are defined here as proactive machines that predict, anticipate, evaluate, and learn from their social context to adapt their behaviour to that context. Ultimately, a socially aware robot must be perceived as trustworthy by humans, a condition that requires transparency, self-explication, and legibility in its behaviour [9,10,11]. However, to be socially aware is to be self-understanding. These robots will require the use of cognitive architectures that allow them not only to navigate and interact in a socially accepted way but also to be able to [12] (i) recognize, understand, and predict the behaviour of people around the robot, to keep a model of their behaviours as well as a semantic representation of the world that allows remembering and planning using social cues; (ii) anticipate not only perceptions but also the consequences of ongoing and possible actions; and (iii) learn to adapt to different situations, contexts, and users, updating their model of others by themselves. Moreover, as a relevant factor for influencing human trust, it is also important that the robot can explain the reason behind its behaviour to the human companion [13], either at their request or on their own initiative because they feel that such explanations are necessary. Causal discovery and reasoning, as one of the current trends in artificial intelligence, is exposed as a potential tool in the management of previously mentioned requirements. From a complementary point of view, as social robots augment their autonomy, enabling them to operate in increasingly open-ended environments, enormous possibilities open for improvements in the human economy and well-being. However, it also poses strong risks that are difficult to assess and control by humans. The recent wave of artificial intelligence based on data-intensive machine learning exacerbates these risks due to (1) severe limitations in the handling of novel situations, and (2) a lack of transparency and explainability [14]. The efficiency, robustness, safety, adaptability, and trust of AI-driven robots are at stake. Suitable technology for the construction of dependable, trustable, autonomous robots is in strong need for both (a) the core operation of the robot, and (b) the engineering process used to build it. Some of these capabilities, resilience, and trust problems of autonomous robots can be traced to robots suffering a lack of understanding of what is going on.
This paper reviews the efforts addressed for allowing a robot to create explanations about its own experience and, in particular, to explain plan updates or perceptual anomalies causally and to integrate these explanations into its body of knowledge and use them later to improve its behaviour. We can set the departure point in the problem of the robot explaining its actions when they are the result of an internal planning process. Those explanations are usually oriented to provide a humanly understandable representation for debugging purposes or to determine liability, or eligibility in autonomous operations [15,16]. However, here we will also review the different problems of providing explanations that can be associated with real communication with the person [17]. This is a very general approach that can be related to scientific and human insight, sometimes stated as the “coordination of claims (i.e., models) and evidence” [18]. Always within the context of the specific goals being pursued and the value that they convey to humans, self-understanding and self-expansion refer to how human–robot communication loops operate [19] and how they can be promoted within a cognitive architecture, making them a structural part of it. Following J. Wyatt, “A system with self-understanding is any system with explicit representations that captures some kind(s) of incompleteness in some kind(s) of knowledge that it possesses. A self-extending system is a system that is able to fill in this incomplete knowledge.” [20].
This paper is organised as follows: Section 2 analyses research efforts carried out in recent years that address the problem of generating explanations to questions that a person can generate for the robot without taking into account or considering constraints. This study considers the inclusion of cause–effect relationships as the common thread in the attempt to make autonomous robots capable of providing explanations and thereby inspire more confidence in the people with whom they share their environment and, sometimes, tasks. The ability of robots to plan how to do a task is not the focus of this survey, but it is closely related to the existence of an internal causal model. The generation of explanations in a scenario where the questions to be answered do not impose conditions is considered first. The robot seeks to explain to the person the reason for the actions it is executing or is going to execute, or the reasons behind a perceptual anomaly. Secondly, we study the case in which the robot needs to give constrained explanations. In this case, the robot needs to revisit the plan and, for instance, determine what happens if instead of one action, another action would have been executed. This involves a process of internalization, which presupposes the existence of a representation and tools that make it possible. Lastly, the issue of self-understanding of perceptual anomalies is addressed. The problem involves reproducing a situation in order to repeat the perceptual experience until, in this emulated environment (imagining), the causes behind the detected anomaly can be determined. In Section 3, the relevant issues are discussed and future lines of research are unfolded. Finally, conclusions are drawn in Section 4.

2. Climbing the Ladder of Causation

In the field of robotics, deep learning techniques and, specifically, deep reinforcement learning (DRL) [14], have been a true revolution, allowing robots to be endowed with both a great perceptive capacity and an advanced ability to learn through the exploration of the environment, with a strategy that the robot itself improves using trial and error techniques [21,22]. However, these techniques rely on models whose complexity prevents people from really understanding how decisions are being made or predictions carried out. To paraphrase Aristotle, we could say that a robot has no knowledge of a thing until it has “grasped its why, that is to say, its cause” [23]. Thus, it can be argued that this lack of ability to offer comprehensive explanations prevents the general acceptance of robotic systems in critical tasks [24]. In their surveys on explainable autonomous robots, Sakai and Nagai [25] and Chakraborti et al. [12] emphasize that, for a robot to be able to have real communication with a person, the robot must not only superficially understand the person’s instructions or questions, but technologies must be developed that allow the person and robot to understand each other’s internal state. As far as the robot is concerned, they state that to do this, the robot must be endowed with the ability to estimate what the person is thinking and to present its ideas in a format that is easy for the person to interpret [26]. As discussed above, if the robot can know the person’s internal state, it will be able to anticipate the person’s needs or adapt its behaviour appropriately. If it can make the person understand what it intends to do through explanations, the robot may truly gain the trust of the people. A robot endowed with these mechanisms is referred to as an explainable autonomous robot (XAR) [9,25]. The concept of XAR is closely linked to that of explainable Artificial Intelligence (XAI), but there is an important difference. XAI methods seek to improve knowledge and control, as well as to provide a framework to justify decisions and predictions [14,27]. However, if this framework focuses more on data (data-driven explainability [25]), in XAR, the problem is typically to explain to humans the actions that the robot itself is autonomously performing in an environment shared with them. In this case, and as Sakai and Nagai [25] rightly point out, the problem is one of goal-driven explainability.
Although they are distinct features of a system, there is an intimate relationship between causality and explainability [13]. In his pioneering work, Lewis defines the act of explaining an event as “providing information about its causal history” [28]. For Tania Lombrozo, an explanation is about filling in the gap that allows the other to understand an event [29]. In this sense, it is a cognitive process, in which answers are given to certain questions, usually of the “why” type and normally subject to certain constraints, since the person receiving the explanation does not want the occurrence of the event per se to be explained, but rather why it occurs in that case and not in other possible (counterfactual) cases [30]. In their causation ladder, J. Pearl and D. Mackenzie set out three steps, which define the three different levels of cognitive ability that a causal learner must possess: seeing, doing, and imagining [31]. The generation of explanations, at each of these levels, will force the robot to possess an increasingly higher cognitive level. The first involves the ability to detect regularities, and irregularities, in the environment. The second involves the ability to predict the effects of a deliberately executed action on the environment. The third allows the agent not only to know the effects caused by an action but also to know why they occur and what to do if the result differs from the expected one. The continued emergence of better sensors, the increased computing power embedded in robots, and the design of new classifiers, many based on deep learning techniques, have meant that, as J. Pearl and D. Mackenzie suggest, robots can now feel comfortably placed on the first step of the causation ladder. Using deep learning for recognising the terrain type, the hexapod robot described in [32] is able to self-adapt in real time to changes in the traversed terrain. In [33], gesture recognition is addressed using a double-size convolution kernel to extract feature information from images, and a two-stream convolutional neural network (two-stream CNN) to recognize the collected gesture instruction data. Figure 1 shows a snapshot of the CLARA robot. This robot can detect and recognise people, and talk to them, and can move safely through the environment, detecting obstacles and avoiding them. The question then becomes, are they able to determine or predict the effects they cause on the environment when they decide to take a deliberate action? Ii addition, if a positive answer can be given to this question, are there proposals that allow a robot to imagine what might have happened if the action taken had been something other than the one executed?
In addition to the activity already mentioned, the three rungs of this causality ladder can be named according to how the agent is able to organize and manage knowledge. Figure 2 is an adaptation of the one proposed by Pearl [35]. Significantly, the figure includes, next to activity and strategies for organising or managing knowledge, the questions that can typically be answered at each step. The first of the steps is called Association, as it only invokes purely statistical relationships. The questions that the robot might ask do not require causal information. Given readings from a laser, the robot can try to estimate its pose, or the probability of colliding with an obstacle, if it maintains a speed of movement. The success of probabilistic robotics rests, in part, on conditional probability: the probability of a random variable X to take a value x, conditioned on a second random variable Y with value y [36]. The second step, Intervention, involves not only perceiving what something is but also acting in the environment. Thus, in this step, we have expressions such as the probability of the event Y = y given that we act and set the value of X to x and subsequently perceive the event Z equal to z. This ability of the robot to intervene is the starting point for the explanation in the work of Sakai and Nagai [25]. Thus, they establish three types of questions associated with the problem of explanation. The question of “What are you going to do?” appears as the first and basic one since it can be considered as the starting point for the other two: “How are you going to do it?” and “Why are you going to do it?”. The first of these can be answered using causal reasoning and causal chains [37,38]. As discussed above, the response seeks to give liability or eligibility to an autonomous activity [16]. The ”Why?” question is, according to Sakai and Nagai [25], the most complex or difficult to answer, and its answer must also be adapted to the intentions of the person asking it.
In any case, all formulations at the Intervention step respond to explain the decisions that the robot has made based on knowledge of the environment that, temporally, responds to the present time. The robot is seeking to achieve a goal (“what?”), it has a plan to achieve it (“how?”), and the actions that make up this plan respond to current knowledge about the context (“why?”). The justifications that the person may ask the robot for may be due to the person’s lack of knowledge of how the planner that drives the robot’s behaviour works internally or to changes in the plan associated, for example, with perceptual anomalies. However, the person’s questions may force the robot to abstract or explain past actions. If the robot, when it is developing the established plan, modifies it to adapt to changes in the context, the question might now be along the lines of “Why did you do this instead of that?” or, more generally, the person might ask “but what if the state were different?” [14]. Drawing a parallel with the two previous steps, the expressions used at this level respond to the probability that event Y is equal to y would be observed had X been x, given that we actually observed X to be x’, and Y to be y’ [35]. Significantly, we are mainly focusing the problem on what some authors call explainable planning. However, it is not all about actions or tasks. Explanations can also have their origin in an unexpected failure or perception, which has arisen during the execution of the plan and that has changed this [39,40]. In many of these situations, the robot is forced to revisit the plan to determine what might have happened. However, when the robot understands that an unexpected change of state may have been due to a geometrically rooted perceptual anomaly, it may have to resort to an internalisation of the situation to explore the problem, for which the use of simulators may be required to imagine what might have happened [41,42].
The next sections provide further details about different methods of knowledge representation in autonomous robots and how they contribute to explainability, and also implemented approaches for providing an answer to the questions in the Intervention and Counterfactuals steps of the causality ladder. As will be discussed in Section 2.2, in its simplest version, to answer the Intervention-related questions, the robot must be able to draw up a plan to achieve a goal, concatenating actions that are triggered by the fulfilment of preconditions. The robot could use this information to answer the “how?” and “why?” questions. To answer the Counterfactuals-related questions, the robot should understand what should be changed in the input instance to change the result obtained. These are post hoc local explanations [43].

2.1. Knowledge Representations and Explainability

Table 1 summarises some typical examples of knowledge representations used by different authors as a basis for enabling an autonomous robot to generate explanations of its behaviour. The table provides reference to these works, the knowledge representation scheme used, the position we could assign to the proposal in the causality ladder (Figure 2), and the applications in which these proposals are used.
In general, it can be said that much of the work in this field has focused on the aforementioned explainable planning. Therefore, many of the representations seek only to describe the course of action, and use objects or locations as attributes of the tasks to be performed. This is, for example, the case of Gao et al. [44]’s proposal, in which so-called And–Or graphs are used. A simple example is shown in Figure 3. The goal in this work is to represent a task that is carried out collaboratively, and in which the robot generates explanations about the progress of the plan and the next task to be performed. The model and internal parameters are adjusted by hand, although the authors leave open the option of learning from annotated user data. The decision-making space can also be encoded as a graph. Zhang et al. [50] propose to represent the model as a sparse multistep graph where nodes are representative landmarks. These landmarks are scattered (in terms of reachability) across the goal space, and learnt via clustering a latent space built using an auto-encoder with reachability constraints. Similarly, in the proposal by Hu and Nagai [51], the robot learns a graph-structured world model consisting of sparse, multistep transitions. Policy nodes based on reachability estimates in the global space are generated. Thus, planning in a continuous state space is simplified to a graph search problem.
Other schemes, popularly used to represent and decompose a task into lower-level objectives, have also been used in frameworks that consider explainability. This is the case of Behaviour Trees (BTs). In the proposal by Han et al. [45], high-level robot explanations are encoded using BTs, framing them into a set of semantic sets (the goal, subgoals, steps, actions). They provided contributed algorithms to generate robot explanations, specifically giving answers to questions seeking causal information. However, as discussed in more detail in Section 2.2, the advantages offered by Automatic Planning make it very common in the context of explainable planning to work with Planning Domain Definition Language (PDDL). Criticisms of the use of PDDL for symbolic planning, due to difficulties in scaling in a real environment [52], do not correspond to the results provided by different works [53,54]. The use of PDDL allows the plan to be synthesised action by action rather than end-to-end as is the case when using deep learning algorithms, which makes it easier to modify it at runtime when the environment changes. Moreover, the latter allows each action to be available for execution as soon as it is generated, which in turn allows concurrent planning and execution. PDDL is used, for example, for the proposals of Krarup et al. [46] or Lindner and Olz [16], which are cited in Table 1, but also for Sreedharan et al. [55] or Stulp et al. [56]. In general, the coding of preconditions and postconditions used in PDDL constitutes a causal model that can be quite complex and that can allow work on reconciliation between the robot and person models [55], or on post hoc explanation [46] (counterfactuals, i.e., an explanation that describes that the output of the model will be y instead of y, thereby changing, without this change occurring in the real world, the behaviour or inputs x to x ) [57].
To address the problem of explaining the reasons behind functional failure, Diehl and Ramirez-Amaro [47] propose the learning of a Bayesian causal network from simulations. The model is then successfully transferred to real environments. The generation of explanations is based on setting the failure state in contrast with the closest state that would have allowed for successful execution. The representation is scalable to manage other problems or platforms.
The cognitive and social sciences have long studied how people explain their own behaviour to others. This research has served as the basis for proposing explanatory frameworks in robotics, which thus emphasise the social-interactive and human behavioural aspects of explanation. Stange et al. [9] propose a framework where several types of memories coexist (working memory, episodic memory). A state machine is encoded as an Interaction model, allowing it to recognize and keep track of the human–robot interaction context. The aim is to give the robot the ability to automatically generate verbal explanations of its own behaviour, mapping its internal courses of action for decision making.
When trying to capture the whole reality of a scenario, without limiting it to a specific application or restricting it only to task management, the knowledge representation must be able to combine different types of tasks, robots, objects, environments, etc. A viable strategy to consider all this information is the use of knowledge-enabled approaches based on ontology [58]. These ontology-based approaches can be adapted to deal with evolving and changing scenarios and different applications, facilitated by the fact that pieces of knowledge, which are in many cases independent of each other, can be reused. There are different semantic knowledge representations in robotics: RoboEarth [59], KnowRob [41], and RoboCSE [60]. The KnowRob architecture allows the robot to emulate a situation, so that it can draw up a plan of action that it can then execute in reality. Although it does not specifically have a module for generating explanations, its internal functioning based on questions and answers, which the robot asks itself to a certain extent, is very interesting. The ARE-OCRA [48] is a framework for generating explanations about target queries. The knowledge base is built using an Ontology for Collaborative Robotics and Adaptation (OCRA).
Some representations are specifically designed to manage knowledge in the application framework addressed by the proposal. For instance, the IID schema proposed by Hanheide et al. [49] stores knowledge about the state and how an action modifies it. By state is meant specific locations and categories of objects, rooms, and places in the building, as well as the whereabouts of particular people and what they know. It organises this knowledge in three layers: instance, default, and diagnostic. The instance layer contains state information. The default layer manages general knowledge about categories (e.g., objects typically found in office rooms) and about the effects of an action on the instance layer. The diagnostic layer contains the action knowledge that is responsible for creating new general knowledge and augmenting the action models in the default layer with possible causes of failure. This knowledge representation is the core of a cognitive architecture, where the flow of information is controlled by employing a distributed blackboard. Task-related agents (dialogue, mapping) are distributed within this architecture in a lower layer. A second layer is included for generating the symbolic items in the knowledge representation. These symbols are used for decision making in the higher layer of the architecture. The existence of a central blackboard as a working memory extended with an internal simulator is the core of the proposal in Bustos et al. [42]. The aim is to endow robots with the ability to navigate and follow those humans being assisted by them while verifying the coherence between the perceived reality and the simulated one. Inconsistencies are detected triggering the search for causal explanations and the acquisition of new concepts.
There are far fewer proposals to address the problem of reasoning, or providing explanations, for a plan that has not actually been implemented. Gjærum et al. [14] propose to use Linear Model Trees (LMTs) for this purpose. The idea is to obtain real-time counterfactual explanations for a black-box model using the LMT as a surrogate model. Figure 4 shows how this scheme would work. Basically, the controller captures the state of the system and returns the appropriate action to execute. The LMT captures that same system state and computes what the counterfactual state would be, which is provided to the controller. With this input, the controller provides a counterfactual action. The combination of state and counterfactual action provides the explanation. Given the low level of the response (the result of a state–action correlation), this system applies to equally low-level controllers.

2.2. Explaining a Plan: How and Why Am I Going to Do It?

Current robotics solutions need to deal with a dichotomic situation: on the one hand, it is expected that they can operate in open-ended environments, but, on the other hand, the complex tasks that they must fulfil are typically specified using static Finite-State Machines or Behaviour Trees. Thus, they are “mostly” blind concerning the future; they usually ignore the impact of the selected state on the next reasoning processes. Moreover, in their standard implementations, they are free-form and static solutions to task planning [45]. As discussed in the previous section, automated solutions do exist to allow for performance planning, rather than mere imitation. Can a robot predict the effects of a deliberatively executed action? Automated Planning [61] is particularly suitable in robotic systems because it employs some form of forward projection (reasoning in depth about goals, preconditions, resources, and timing constraints) and provides the robot reasonable responses in states unforeseen by the designer [62,63]. The comparison of the behaviour of a planner in a scenario where the robot shares its environment, and perhaps task, with people, will be based on objective measures (such as efficiency in task execution, accuracy or time spent), and subjective measures, which can be captured with Likert-scale questionnaires [64]. These questionnaires will assess determinants such as situational awareness, complacency, likability, trust, or mental workload. The aim of providing explanations should be to improve these metrics. The topic is addressed in detail in Krarup et al. [46]’s work.
Typically stated as a Markov Decision Process, a decision-making space Π can be expressed using the following transition function [12,25]
δ Π : A × S S × ,
where Π is the decision-making space, A and S are the actions available to the robot and the set of possible states it can be in, respectively, and represents the cost (expressed as a real number) required for the transition. Using this terminology, the plan π is a sequence of actions
π = { a 1 , a 2 , . . . a i } , a i A ,
which is generated using an algorithm A and under a certain constraint τ , i.e.
A : Π × τ π ,
Briefly, the plan transforms the current state X S to its goal G S , i.e.,
δ Π ( τ , X ) = G , a i π c i ,
where a i π c i denotes the plan cost c ( π ) [40]. The policy can be expressed as
π : s a , a A , s S .
Explanations typically justify to the person that the solution π satisfies τ for a given Π . Correctly, Das et al. [40] add a new explanation type: that which allows the person to understand that a wrong state has been reached unexpectedly, which has meant stopping the execution of the plan π . In any case, the most basic explanation of a plan consists of the planner explaining the decisions it has made based on its internal model (i.e., current state, actions, and goals) [16,17]. In the Cambridge Dictionary, this form of expressing oneself is called soliloquy, and is defined as “a speech in a play that the character speaks to himself or herself or the people watching rather than to the other characters”. Obviously, these explanations will only be valid if the person receiving them is intimately familiar with the internal model of the robot, with how the robot organizes knowledge, and how the planner decomposes the execution of a plan. Otherwise, when the human interlocutor is a non-expert user, possessing a domain and task model that differs significantly from those internalized by the robot, the validity of these explanations may be far from being of any use. Figure 5 schematizes this situation. The terminology used follows that proposed by Sakai and Nagai [25]. Both the robot and the person are carrying out their plans. Assuming that the decision-making process can be applied to people and robots, we can express these internally generated policies as
A ¯ H H : Π ¯ H H × τ ¯ H H π ¯ H H A ¯ R R : Π ¯ R R × τ ¯ R R π ¯ R R
where A ¯ H H refers to the person’s cognitive model of its own needs, behaviours, goals, etc., and A ¯ R R to the robot’s computational model of its own behaviours, goals, etc. Being internally generated, they are not directly observable by the others. The interpretable policies can be expressed as
A H H : Π H H × τ H H π H H A R R : Π R R × τ R R π R R
Internally, the person interacts with a model of the robot, A R H . This interaction process can be expressed as
A H H : Π H H × τ H H π H H A R H : Π R H × τ R H π R H
For this to be possible, the plan generated by the robot concerning its model needs to be interpreted by the person ( Π R H ). However, as it is shown in Figure 5, if the robot does not provide or provides insufficient explanations (or its behaviour does not make sense with respect to the person’s model), this will be interpreted by the person with respect to her own model ( Π H H ). Although we consider in the figure that the human understands the robot’s constraints (they are an estimate, τ R H , of τ R R ), further explanations are needed.
Sakai and Nagai [25] identify four requirements for allowing a robot to generate an explanation. Briefly, the internal model of the robot must be interpretable by the person, thus avoiding the fact that the robot behaves strangely for the person, and the person and the robot must have adequate versions of the elements that define the other’s plan (model, constraints, plan generation algorithm). Finally, the explanations provided by the robot must be understandable to the person, avoiding verbalization or visualization problems. Figure 5 shows the problem that arises when one, or more, of these requirements are not met. The result shown in the figure may be the result of a robot whose behaviour is not reasonable or understandable to the person (it may be caused by the internal model used by the robot). However, the problem could also be that the internal models used by the robot and person, Π ¯ R R and Π ¯ H H , respectively, are very different (actions, states, etc.). This is more common when the physical capabilities of the robot are very different from those of the person and, as aforementioned, when the person is a non-expert [40,65]. To deal with this situation, Raman and Kress-Gazit [65] propose to use Linear Temporal Logic. However, in this proposal, explanations focus on which actions are infeasible for the robot but not on why they are not. Therefore, these explanations are often insufficient. Das et al. [40] propose a new type of explanation, which describes the cause of unexpected failures during the execution of the plan. They show that explanations capturing the context of a failure and also the history of past actions are the most effective for failure and solution identification among non-experts. The four requirements for achieving an explainable robotic agent are summarized in Table 2.
The first requirement implies that the robot must internalise a model that is interpretable by the person. The goal is to provide good explanations for the person, not for the robot itself. However, several approaches focus on providing good explanations from the perspective of the robot [49]. To alleviate this problem, some work actually assumes that the models for the person and the robot can be the same. In this case, an explanation from the robot’s perspective would not only be correct but also perfectly understandable for the human [75]. As the references shown in Table 2 show, it is very common to employ hierarchical models to represent the plan (And–Or graphs, Behaviour Trees) which, in collaborative frameworks, encode both the one maintained by the robot and the one inferred by the person. For instance, in the proposal by Gao et al. [44], a spatial–temporal–causal And–Or graph is employed for encoding the joint plan (Figure 3). Significantly, this graph encodes a joint task plan, providing a unified representation of the robot’s knowledge and plan as well as the inferred person’s knowledge and plan. During the collaboration, the robot manages an instance of this graph to represent its internal state and another one to represent the inferred person’s internal state. They share the same structure, and, at a given time t, the robot can compare the most likely person’s mental state with its internal state. If they are the same, there is no need to explain to the person. In the other case, the robot should provide the explanation. As described in Section 2.1, the decision-making space can also be encoded as a graph [50,51].
However, in a more general case, it is assumed that both models differ. In the context of Automated Planning and Robotic systems, Problem Modelling Languages such as the aforementioned PDDL [76,77] constitute an attempt to standardize the knowledge representation of planning problems using a high level of abstraction. They allow us to model the physical states, the transformations between these states, and the goals and the requirements of the robot tasks in a declarative form, regardless of the robotic platform or the specific programming language used in the low-level modules of the architecture. The formulation of planning as a state transition system has been at the core of domain-independent planning research and, specifically, of PDDL. Thus, although several extensions have been devised to extend the expressiveness of PDDL [78], the focus remains on handling causal relationships derived from the conditions and effects of primitive actions. Equation (11) was initially proposed by Chakraborti et al. [17] as a model reconciliation problem (MRP), i.e., the problem of bringing the human’s model closer to the robot’s model by making minimal changes. In their work, PDDL was used to state and analyse the MRP. Given the high-level description in PDDL of the domain, the problem and the goals to be reached, the planning systems aim to find a sequence of actions to be executed by the robot to reach those goals. In this case, the planning system provides a flow of actions that is adapted in runtime to the changes in the environment. Regarding the latter, contextual information can be considered as part of the state description of the planning system. Depending on the runtime estimation of these context-related metrics, the planning system could decide to adopt the nominal course of action. To implement this idea, the planning system should manage in a high-level of abstraction the execution of the low-level actions, which will be distributed in the task-related modules of the robotic architecture. Therefore, it can be noted that the use of Automated Planning in robot control will require the definition of at least two levels of abstraction: high-level (deliberative planning) and low-level (robot sensing and reactive behaviours). The mapping between these two levels of abstraction is required to solve real tasks [79,80,81]. Typically, such mappings are written in a specific programming language by experts who know the software, in an ad hoc manner for each particular robot platform and task planning system. In this way, including modifications requires, on the one hand, to resort to experts who know the software architecture and, on the other hand, rewriting the source code, and recompiling the entire software architecture in the worst case. However, ideally, such mapping could also be defined in a declarative way, so that they can be easily edited (even by non-experts) and without modifying the modules of the control architecture, which should be able to process such a formal description.

2.3. Exploring the Recent Past: Why Did You Do This Instead of That?

As reviewed in the previous section, the need for an explanation typically arises when a mismatch is detected between the plan being executed by the robot and the one expected by the person. To try to find out where the difference between the plans lies, the person expresses his or her doubts with questions. However, several studies [14,43,46] have shown that, when people ask questions about plans, those questions are contrastive, i.e., ”Why did you do this instead of that?”. The difference between this question and the question “Why do you do this?” is relevant, as it involves removing ambiguity to some extent. A contrastive question aims to include the context of the question, providing a more realistic idea of what the questioner needs in the explanation. For instance, in Krarup et al. [46]’s proposal, this situation is posed as a plan negotiation problem. Thus, in order to get the person to understand the robot’s plan, an iterative process of questions and answers (explanations) arises, the ultimate goal of which is that the person is able to understand this plan π R R . A domain-independent approach is proposed to compile these questions into constraints. These constraints are added to the planning model, so that a solution to the new model represents the contrastive plan that is used to produce an explanation.
Contrastive questions can be considered as local or global [82]. Local questions are asked when the person wants an explanation of specific decisions made by the robot at a given time, while global questions are asked when one wants to better understand how the robot makes decisions in general. In both cases, they can be restricted to a specific model. In one case, questions are asked to find out why a decision has been made (or what could be expected if the decision had been different), while in the global case, questions seek knowledge about the model itself. In this paper, we will focus on the more local framework: the person has a relatively deep knowledge of the model, and the questions are asked to find out details about the execution of the plan or to corroborate that the plan was adequate (or where it might have gone wrong). However, this is not to say that many of the methods discussed in this study are not capable of being employed in different scenarios. In the specific case of the method proposed by Krarup et al. [46], the idea is that the approach implemented to compile the issues into plan constraints (and thus solve the plan negotiation problem) is domain-independent. The example described in the article could easily be a different one. The same is true for the Linear Model Trees approach proposed by Gjærum et al. [14], which is in fact applied in this work in two very different scenarios.
Counterfactual explanations attempt to answer hypothetical questions, which involve revisiting the recent past to analyse states or actions [83]. Although they could be posed in the form: “But what would you have done if the state had been different?”, the question typically takes the form of “Why did you do this instead of that?”. In the typical examples used to describe what a counterfactual explanation is, the emphasis is on the importance of providing actionable explanations, i.e., explanations that would enable the recipient to change certain parameters and thereby successfully achieve the desired outcome. This is certainly the ultimate goal of explaining. However, as Gjærum et al. [14] rightly describes, in Robotics, it does not make much sense for an explanation to seek that you change the input parameters (the state), but rather that they are feasible, i.e., that both the counterfactual state and the counterfactual action are physically possible [43].
To provide these post hoc explanations, Carreira-Perpiñán and Hada [84] propose a method based on classification trees with both univariate and multivariate splits in the branch nodes. Karimi et al. [85] employ a method in which finding these explanations is expressed as a sequence satisfiability problem. Neither of these two solutions can be applied to regression problems with multiple outputs, which makes it difficult to apply them to the complex scenarios in which a real human–robot interaction process will be framed. The concept of the leaf-to-leaf counterfactual distance matrix is proposed by Sokol and Flach [86] to describe how much an instance belonging to a leaf node must be changed to belong to a different node. The framework is again classification trees. Gjærum et al. [14] propose to use decision trees with linear functions in the leaf nodes (linear model trees [87]), showing that they can be valid for generating counterfactual explanations in relatively competitive scenarios with multiple, continuous outputs. Specifically, the challenge is solved using H-LMTs (Heuristic Linear Model Trees) [88]. An H-LMT does not take maximum depth as a restriction but rather the maximum number of leaf nodes allowed. This allows the growing of asymmetric trees that grow deeper in the more complex parts of the feature space while leaving the parts of the tree covering simpler regions shallow. Krarup et al. [46] set up the problem of generating an explanation as an iterative process, in which the person repeatedly asks the robot. These questions are contrastive in nature, and each of them leads to a more constrained planning problem (e.g., the question “Why did you do A instead of B?” involves drawing a plan in which B is in place of A). Solving this new planning problem allows the person to compare and contrast the alternative proposed by the person with respect to the original plan, and allows the person to gain knowledge about the robot’s internal model and the robot to gain knowledge about the person’s preferences. This proposal formalises the so-called explanation as exploration, based on the idea that the person interested in learning more about how a black-box model works will ask contrastive questions in an iterative way [89].

2.4. (Self)Explaining Perceptual Anomalies

Elaborating on the previous section, a more specific need for explanation arises when the robot experiences a perceptual anomaly. An anomaly can be defined as a substantial deviation in the perceptual input from a model-based prediction. For instance, we can imagine a scenario in which a robot follows a human, carrying some object for her. At some point, the robot hits an unforeseen pothole and the object tips over. We assume that both the tilting caused by the pothole and the falling of the box can be detected by its sensors, let us say an Inertial Measurement Unit (IMU) and a camera.
This scenario provides the basic ingredients to trigger the search for a causal explanation and, more crucially, the use of the explanation to make structural changes in the robot’s internal model, providing a potential way for learning a new concept, a new perceptual routine, or a new behaviour. Additionally, this causal explanation could be transformed into a natural language explanation for the surprised human that is leading the way. To be effective, this transformation should take into account the compatibility between models discussed earlier.
As suggested by Bustos et al. [42], the self-explanation of perceptual anomalies can be approached using an internal simulator as part of the robot’s cognitive architecture. The use of internal simulators has gained momentum in recent years. Several research groups are including them in their cognitive robot architectures to explore, among other things, the trade-off between potential benefits and software complexity [90,91,92,93]. These simulators have been used to generate predictions, select anticipatory actions, solve occlusion problems in object or human tracking, apply basic physical laws to perceived objects to stabilise them, or interpret human movements by projecting their kinematics into the robot’s own model. A new potential use of internal simulators is their role as hypothesis testers in the construction of contrastive causal explanations. Following the scenario presented in Figure 6, the internal simulator in the robot would be used in two different ways. First, it would run in synchrony with the robot’s perception and control system to generate short-term predictions of the perceived world. This mode is exactly what is needed to detect the perceptual anomaly caused by the bump in the first place. Since the robot did not see the bump, its internal simulator did not reflect it, and its predictions differed from the real IMU and camera measurements. This difference triggers the search for a cause that explains the anomaly and the new situation where the object is no longer on the tray. The second way of using the internal simulator will become clear in what follows. The search process could be divided into the following steps:
  • The conversion of the anomaly data to a formal representation.
  • The search for a set of potential hypotheses.
  • The conversion of each hypothesis into an initial state of the inner simulator
  • The replay of the duration of the event since the detection of the anomaly starting with each selected hypothesis, and possible sweeping its internal parameters (i.e., position of the bump and width, depth, or height)
  • The comparison in each run of the synthetic and real sensorial data (i.e, IMU) to check that the timings are an approximate match
  • The selection of the best hypothesis and conversion to a formal representation, which can be used to modify and extend the internal knowledge of the robot and to generate a natural language explanation.
As this introspective sequence begins, the internal simulator is used as a replay machine, feeding from an episodic memory to reproduce the past event, starting from the occurrence of the anomaly. The scheme resembles the narrative-enabled episodic memories (NEEMs) proposed by Beetz et al. [41]. NEEMs are constructed with an experience and a narrative. The NEEM experience stores the highly detailed recording of an experience (postures, perceptions, etc.), while the NEEM narrative is described as a story, with a more abstract or symbolic content. In [42]’s proposal, episodic memory shares geometric and symbolic elements in the same graph. As in KnowRobSIM [94], the geometric part is manipulated using a game engine-based system. The software architecture of the game engine-enabled knowledge processing is schematised in Figure 7. Briefly, the inner world performs a loop in which (1) the dynamic state of the agent and world is updated (e.g., with control signals sent to the joint motors of the robot), (2) the world state is updated considering the control inputs generated by the agent (physics are taken into account), and (3) the updated world state is visually rendered.
Returning to the example in Figure 6, the search for a set of plausible hypotheses could be based on an existing ontology and semantic memory, an external query to a wider resource, or even a prompt to an internal Large Language Model (LLM) [95]. With these initial conditions, the simulator can be used online to test whether the new potential causes can generate the sequence of sensory events previously registered. If so, the robot could conclude that it has found a plausible causal explanation for what happened and use it to improve its future behaviour and to communicate with the human. These two subsequent actions would eventually generate a confirmation or refutation of the explanation, feeding this learning process.
As this imagined scenario attempts to show, combining detectable perceptual anomalies with appropriate processes that can construct causal explanations, based on the acquisition and testing of a set of contrastive hypotheses, could be a promising path towards a self-explaining robotic cognitive architecture. Although there are many open problems to be solved, this approach is based from the outset on methods that will contribute to self-explaining robots.

3. Discussion and Future Research Work

The last few decades have seen a true revolution in the perceptual capabilities that an autonomous robot can deploy. Algorithms that are already considered traditional have been practically replaced by approaches based on Deep Learning. Along with the evolution of algorithms, hardware and application scenarios have matured, so that, today, the robot can perceive the surrounding environment and build a robust and reliable semantic representation. Based on this perceptive capacity, and making use of algorithms based on the probabilistic approaches of the beginning of the century, autonomous robots are capable of solving certain tasks, such as navigation or localisation, in a robust manner. Equipped with all these hardware and software apparatus, autonomous robots are nowadays deployed in scenarios where they share space with people. It is not difficult to see them working as waiters in restaurants, or handling pallets in large intralogistics warehouses. However, deployment in scenarios where the nature of the tasks to be tackled, or the profile of the people with whom the environment is shared, means that the robot’s behaviour is not always correctly interpreted by these people, forcing the robot to have to explain its behaviour. The field of work on the development of autonomous explainable robots is relatively young but has been gaining momentum in recent years. As Figure 8 shows, the number of publications considering the topics Explanation and Robot has grown exponentially since the beginning of the century (data are taken from the Web of Science). Although the data for 2023 are incomplete (the data capture date in the figure is January 2024), in recent years, it has exceeded more than 100 publications per year, whereas in 2000, it was less than 20.
Efforts in explanation generation have focused primarily on environments where robots and humans work together to solve a given task. In these cases, it is even possible for both actors to internalise the same causal model, which facilitates the detection of disparities. These disparities or inconsistencies will be the basis for the generation of explanations by the robot, always aimed at ensuring that the joint plan runs satisfactorily. In this same direction, the generation of explanations is linked to the existence of this plan, which the robot knows or learns, and which constitutes the reference that the robot seeks to be fulfilled. Adding constraints to the planning problem, with questions aimed, for example, at asking the robot to explain the reasons why one activity was executed instead of another, is framed as a restricted planning problem, which can be solved as the original unconstrained one and with a similar time.
Future work will need to address different issues. As discussed above, when task execution is shared between robots and humans, disparities arise from the side-by-side comparison of plans. This scheme does not consider the plans’ causal structure to generate the explanations, nor does it go into details that might require abstracting in some way from the predicates, variables, and/or actions. It is also relevant to consider the problem of unsolvability, i.e., the explanation of solutions or problems that complicate or impede the resolution of the task. In this sense, the use of solutions that consider geometric emulation (including, for example, physical models) in order to establish causal relationships at the level of the evolution of the context and not only of the action, can be helpful. An alternative approach uses PDDL to express planning problems and has important advantages, but it does not have expressiveness for considering constraints on inclusion, exclusion, or ordering actions, nor does it allow more complex constraints on the structure of the plan or on how something is achieved to be easily expressed (e.g., “Why did you use action A rather than action B for achieving P?”). To deal with this issue, instead of expressing constraints over the state trajectory of the plan (as the Linear Temporal Logic over finite traces [96] or the PDDL3 [97] suggest), plan constraints can be expressed over trajectories of actions [98]. This will allow performance to be improved by exploiting knowledge expressed by action constraints [46]. Additionally, there are less popular solutions that will also need to be explored. One example is Answer Set Programming (ASP), designed for knowledge-intensive reasoning but also used in task planning problems. In the work by Jiang et al. [99], ASP-based planners outperformed PDDL-based planners when the number of objects and tasks was large, and thus, reasoning about complex precondition relationships and action effects was relevant.
The problem of explanation generation can also be approached from a slightly different angle: the generation of explanations triggered by perceptual anomalies and the translation of these explanations into a human-understandable format. These anomalies are likely to occur frequently in the daily operation of a social robot. Internal simulators can be an effective way to test contrastive hypotheses online, but several problems remain to be solved. One of the most challenging is the online transfer between the numerical representations needed to run simulations and the symbolic ones needed to search and extend the robot’s internal knowledge. This coming and going of the transformations required to build a new causal explanation is, from the outset, a promising place to pursue research in the creation of new concepts.

4. Conclusions

This article describes the general framework of the generation of explanations by an autonomous robot, starting from the basic concept of its internalization of a more or less complex scheme of cause–effect relationships. The review of the state of the art allows us to assess how, in recent years and after the great evolutionary leap experienced in the perceptual abilities of these robots, the integration of causality has begun to be considered as one of the basic pillars of the creation of more acceptable robots that can be considered by people as credible, safe, and reliable. Specifically, this causality serves as the basis for the robot to generate explanations, either limited to a temporal sequence that is being traversed (with a series of specific actions, already executed or to be executed in the future) or open to the analysis of temporal sequences not initially considered (in which there are actions that could have been executed but have not been, or in which perceptions have been disregarded, which may be the cause of unforeseen results).

Author Contributions

Conceptualization, R.M., P.B. and A.B.; Methodology, R.M., P.B. and A.B.; Formal analysis, R.M., P.B. and A.B.; Investigation, R.M., P.B. and A.B.; Writing—original draft, R.M., P.B. and A.B.; Funding acquisition, R.M., P.B. and A.B. All authors contributed equally to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been supported by grants PDC2022-133597-C4X, TED2021-131739B-C2X, and PID2022-137344OB-C3X, funded by MCIN/AEI/10.13039/501100011033 and by the European Union NextGenerationEU/PRTR (for the first two grants), and “ERDF A way of making Europe” (for the third grant).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gladden, M.E. Who Will Be the Members of Society 5.0? Towards an Anthropology of Technologically Posthumanized Future Societies. Soc. Sci. 2019, 8, 148. [Google Scholar] [CrossRef]
  2. Matarić, M.J. Socially assistive robotics: Human augmentation versus automation. Sci. Robot. 2017, 2, eaam5410. [Google Scholar] [CrossRef] [PubMed]
  3. Sgorbissa, A.; Papadopoulos, I.; Bruno, B.; Koulouglioti, C.; Recchiuto, C. Encoding Guidelines for a Culturally Competent Robot for Elderly Care. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1988–1995. [Google Scholar] [CrossRef]
  4. Seibt, J.; Damholdt, M.; Vestergaard, C. Integrative Social Robotics, value-driven design, and transdisciplinarity. Interact. Stud. 2020, 21, 111–144. [Google Scholar] [CrossRef]
  5. Tian, L.; Oviatt, S. A Taxonomy of Social Errors in Human-Robot Interaction. ACM Trans. Hum.-Robot Interact. 2021, 10, 1–32. [Google Scholar] [CrossRef]
  6. Rossi, S.; Rossi, A.; Dautenhahn, K. The Secret Life of Robots: Perspectives and Challenges for Robot’s Behaviours During Non-interactive Tasks. Int. J. Soc. Robot. 2020, 12, 1265–1278. [Google Scholar] [CrossRef]
  7. Sandini, G.; Sciutti, A.; Vernon, D. Cognitive Robotics. In Encyclopedia of Robotics; Ang, M.H., Khatib, O., Siciliano, B., Eds.; Springer: Berlin/Heidelberg, Germany, 2020; pp. 1–7. [Google Scholar] [CrossRef]
  8. Hellström, T.; Bensch, S. Understandable robots—What, Why, and How. Paladyn J. Behav. Robot. 2018, 9, 110–123. [Google Scholar] [CrossRef]
  9. Stange, S.; Hassan, T.; Schröder, F.; Konkol, J.; Kopp, S. Self-Explaining Social Robots: An Explainable Behavior Generation Architecture for Human-Robot Interaction. Front. Artif. Intell. 2022, 5, 866920. [Google Scholar] [CrossRef] [PubMed]
  10. Sciutti, A.; Barros, P.; Castellano, G.; Nagai, Y. Editorial: Affective shared perception. Front. Integr. Neurosci. 2022, 16, 1024267. [Google Scholar] [CrossRef] [PubMed]
  11. Woitsch, R.; Utz, W.; Sumereder, A.; Dieber, B.; Breiling, B.; Crompton, L.; Funk, M.; Bruckmüller, K.; Schumann, S. Collaborative Model-Based Process Assessment for Trustworthy AI in Robotic Platforms. In Proceedings of the International Conference on Society 5.0, Online, 22–24 June 2021; pp. 163–174. [Google Scholar] [CrossRef]
  12. Chakraborti, T.; Sreedharan, S.; Kambhampati, S. The Emerging Landscape of Explainable Automated Planning & Decision Making. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI’20, Yokohama, Japan, 7–15 January 2021. [Google Scholar]
  13. Setchi, R.; Dehkordi, M.B.; Khan, J.S. Explainable Robotics in Human-Robot Interactions. Procedia Comput. Sci. 2020, 176, 3057–3066. [Google Scholar] [CrossRef]
  14. Gjærum, V.B.; Strümke, I.; Lekkas, A.M.; Miller, T. Real-Time Counterfactual Explanations For Robotic Systems With Multiple Continuous Outputs. IFAC-PapersOnLine 2023, 56, 7–12. [Google Scholar] [CrossRef]
  15. Rodríguez-Lera, F.; González-Santamarta, M.A.; Guerrero-Higueras, A.; Martín, F.; Matellán, V. Towards explainability in robotics: A performance analysis of a cloud accountability system. Expert Syst. 2022, 39, e13004. [Google Scholar] [CrossRef]
  16. Lindner, F.; Olz, C. Step-by-Step Task Plan Explanations Beyond Causal Links. In Proceedings of the 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), Napoli, Italy, 29 August–2 September 2022; pp. 45–51. [Google Scholar] [CrossRef]
  17. Chakraborti, T.; Sreedharan, S.; Zhang, Y.; Kambhampati, S. Plan Explanations as Model Reconciliation: Moving Beyond Explanation as Soliloquy. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 156–163. [Google Scholar] [CrossRef]
  18. Sandoval, W.A. Epistemic Goals. In Encyclopedia of Science Education; Gunstone, R., Ed.; Springer: Dordrecht, The Netherlands, 2015; pp. 393–398. [Google Scholar] [CrossRef]
  19. Kontogiorgos, D. Utilising Explanations to Mitigate Robot Conversational Failures. arXiv 2023, arXiv:cs.HC/2307.04462. [Google Scholar]
  20. Wyatt, J.L.; Aydemir, A.; Brenner, M.; Hanheide, M.; Hawes, N.; Jensfelt, P.; Kristan, M.; Kruijff, G.J.M.; Lison, P.; Pronobis, A.; et al. Self-Understanding and Self-Extension: A Systems and Representational Approach. IEEE Trans. Auton. Ment. Dev. 2010, 2, 282–303. [Google Scholar] [CrossRef]
  21. Remman, S.B.; Lekkas, A.M. Robotic Lever Manipulation using Hindsight Experience Replay and Shapley Additive Explanations. In Proceedings of the 2021 European Control Conference (ECC), Online, 29 June–2 July 2021; pp. 586–593. [Google Scholar] [CrossRef]
  22. Haarnoja, T.; Ha, S.; Zhou, A.; Tan, J.; Tucker, G.; Levine, S. Learning to walk via deep reinforcement learning. arXiv 2018, arXiv:1812.11103. [Google Scholar]
  23. Falcon, A. Aristotle on Causality. In The Stanford Encyclopedia of Philosophy, Spring 2023 ed.; Zalta, E.N., Nodelman, U., Eds.; Metaphysics Research Lab, Stanford University: Stanford, CA, USA, 2023. [Google Scholar]
  24. Edmonds, M.; Gao, F.; Liu, H.; Xie, X.; Qi, S.; Rothrock, B.; Zhu, Y.; Wu, Y.N.; Lu, H.; Zhu, S.C. A tale of two explanations: Enhancing human trust by explaining robot behavior. Sci. Robot. 2019, 4, eaay4663. [Google Scholar] [CrossRef] [PubMed]
  25. Sakai, T.; Nagai, T. Explainable autonomous robots: A survey and perspective. Adv. Robot. 2022, 36, 219–238. [Google Scholar] [CrossRef]
  26. Zakershahrak, M.; Gong, Z.; Zhang, Y. Online Explanation Generation for Human-Robot Teaming. arXiv 2019, arXiv:abs/1903.06418. [Google Scholar]
  27. Adadi, A.; Berrada, M. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). IEEE Access 2018, 6, 52138–52160. [Google Scholar] [CrossRef]
  28. Lewis, D. Causal Explanation. In Philosophical Papers Vol. Ii; Lewis, D., Ed.; Oxford University Press: Oxford, UK, 1986; pp. 214–240. [Google Scholar]
  29. Lombrozo, T. The structure and function of explanations. Trends Cogn. Sci. 2006, 10, 464–470. [Google Scholar] [CrossRef] [PubMed]
  30. Matarese, M.; Rea, F.; Sciutti, A. A User-Centred Framework for Explainable Artificial Intelligence in Human-Robot Interaction. arXiv 2021, arXiv:cs.AI/2109.12912. [Google Scholar]
  31. Pearl, J.; Mackenzie, D. The Book of Why; Basic Books: New York, NJ, USA, 2018. [Google Scholar]
  32. Chen, G.; Han, Y.; Li, Y.; Shen, J.; Tu, J.; Yu, Z.; Zhang, J.; Cheng, H.; Zhu, L.; Dong, F. Autonomous gait switching method and experiments of a hexapod walking robot for Mars environment with multiple terrains. Intell. Serv. Robot. 2024, 17, 1–21. [Google Scholar] [CrossRef]
  33. Zhang, C.; Li, H. Adoption of Artificial Intelligence Along with Gesture Interactive Robot in Musical Perception Education Based on Deep Learning Method. Int. J. Humanoid Robot. 2022, 19, 2240008. [Google Scholar] [CrossRef]
  34. Romero-Garcés, A.; Bandera, J.P.; Marfil, R.; González-García, M.; Bandera, A. CLARA: Building a Socially Assistive Robot to Interact with Elderly People. Designs 2022, 6, 125. [Google Scholar] [CrossRef]
  35. Pearl, J. Theoretical Impediments to Machine Learning With Seven Sparks from the Causal Revolution. arXiv 2018, arXiv:cs.LG/1801.04016. [Google Scholar]
  36. Thrun, S.; Burgard, W.; Fox, D. Probabilistic Robotics; MIT Press: Cambridge, MA, USA, 2005. [Google Scholar]
  37. Lagnado, D.; Channon, S. Judgments of cause and blame: The effects of intentionality and foreseeability. Cognition 2008, 108, 754–770. [Google Scholar] [CrossRef]
  38. Hilton, D.J.; McClure, J.; Sutton, R.M. Selecting explanations from causal chains: Do statistical principles explain preferences for voluntary causes? Eur. J. Soc. Psychol. 2010, 40, 383–400. [Google Scholar] [CrossRef]
  39. Khanna, P.; Yadollahi, E.; Björkman, M.; Leite, I.; Smith, C. Effects of Explanation Strategies to Resolve Failures in Human-Robot Collaboration. In Proceedings of the 2023 32nd IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), Busan, Republic of Korea, 28–31 August 2023; pp. 1829–1836. [Google Scholar] [CrossRef]
  40. Das, D.; Banerjee, S.; Chernova, S. Explainable AI for Robot Failures: Generating Explanations That Improve User Assistance in Fault Recovery. In Proceedings of the 2021 ACM/IEEE International Conference on Human-Robot Interaction, Boulder, CO, USA, 9–11 March 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 351–360. [Google Scholar] [CrossRef]
  41. Beetz, M.; Beßler, D.; Haidu, A.; Pomarlan, M.; Bozcuoğlu, A.K.; Bartels, G. Know Rob 2.0—A 2nd Generation Knowledge Processing Framework for Cognition-Enabled Robotic Agents. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 512–519. [Google Scholar] [CrossRef]
  42. Bustos, P.; Bachiller, P.; García-Varea, I.; Martínez-Gómez, J.; Bandera, A.; Marfil, R. INSIGHT: The quest for causal explanations. In Proceedings of the WAF2023—23 Workshop on Physical Systems, Aranjuez, Spain, 9–10 November 2023. [Google Scholar]
  43. Guidotti, R. Counterfactual explanations and how to find them: Literature review and benchmarking. Data Min. Knowl. Discov. 2022, 36, 1–55. [Google Scholar] [CrossRef]
  44. Gao, X.; Gong, R.; Zhao, Y.; Wang, S.; Shu, T.; Zhu, S.C. Joint Mind Modeling for Explanation Generation in Complex Human-Robot Collaborative Tasks. In Proceedings of the 2020 29th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), Naples, Italy, 31 August–4 September 2020; pp. 1119–1126. [Google Scholar] [CrossRef]
  45. Han, Z.; Giger, D.; Allspaw, J.; Lee, M.S.; Admoni, H.; Yanco, H.A. Building the Foundation of Robot Explanation Generation Using Behavior Trees. J. Hum.-Robot Interact. 2021, 10, 26. [Google Scholar] [CrossRef]
  46. Krarup, B.; Krivic, S.; Magazzeni, D.; Long, D.; Cashmore, M.; Smith, D.E. Contrastive Explanations of Plans through Model Restrictions. J. Artif. Int. Res. 2021, 72, 533–612. [Google Scholar] [CrossRef]
  47. Diehl, M.; Ramirez-Amaro, K. Why Did I Fail? A Causal-Based Method to Find Explanations for Robot Failures. IEEE Robot. Autom. Lett. 2022, 7, 8925–8932. [Google Scholar] [CrossRef]
  48. Alarcos, A.O.; Foix, S.; Alenyà, G. Knowledge Representation for Explainability in Collaborative Robotics and Adaptation. In Proceedings of the Workshop on Data Meets Applied Ontologies in Explainable AI (DAO-XAI 2021), Part of Bratislava Knowledge, September (BAKS 2021), IAOA, Bratislava, Slovakia, 18–19 September 2021. [Google Scholar]
  49. Hanheide, M.; Göbelbecker, M.; Horn, G.S.; Pronobis, A.; Sjöö, K.; Aydemir, A.; Jensfelt, P.; Gretton, C.; Dearden, R.; Janicek, M.; et al. Robot task planning and explanation in open and uncertain worlds. Artif. Intell. 2017, 247, 119–150. [Google Scholar] [CrossRef]
  50. Zhang, L.; Yang, G.; Stadie, B.C. World Model as a Graph: Learning Latent Landmarks for Planning. arXiv 2020, arXiv:2011.12491. [Google Scholar]
  51. Hu, S.; Nagai, T. Explainable autonomous robots in continuous state space based on graph-structured world model. Adv. Robot. 2023, 37, 1025–1041. [Google Scholar] [CrossRef]
  52. Capitanelli, A.; Mastrogiovanni, F. A Framework for Neurosymbolic Robot Action Planning using Large Language Models. arXiv 2023, arXiv:cs.AI/2303.00438. [Google Scholar]
  53. Pulido, J.C.; González Dorado, J.C.; Suárez, C.; Bandera, A.; Bustos, P.; Fernández, F. Evaluating the Child–Robot Interaction of the NAOTherapist Platform in Pediatric Rehabilitation. Int. J. Soc. Robot. 2017, 9, 343–358. [Google Scholar] [CrossRef]
  54. Li, J.; Cong, M.; Liu, D.; Du, Y. Integrated Classical Planning and Motion Planning for Complex Robot Tasks. In Intelligent Robotics and Applications; Liu, X.J., Nie, Z., Yu, J., Xie, F., Song, R., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 70–78. [Google Scholar]
  55. Sreedharan, S.; Chakraborti, T.; Kambhampati, S. Foundations of explanations as model reconciliation. Artif. Intell. 2021, 301, 103558. [Google Scholar] [CrossRef]
  56. Stulp, F.; Bauer, A.S.; Gomez, S.B.; Lay, F.S.; Schmaus, P.; Teams, D.R.G.B.; Leidner, D. Explainability and Knowledge Representation in Robotics: The Green Button Challenge. In Proceedings of the Explainable Logic-Based Knowledge Representation (XLoKR 2020), Rhodes, Greece, 12–18 September 2020. [Google Scholar]
  57. Verma, S.; Dickerson, J.; Hines, K. Counterfactual Explanations for Machine Learning: Challenges Revisited. arXiv 2021, arXiv:cs.LG/2106.07756. [Google Scholar]
  58. Manzoor, S.; Rocha, Y.G.; Joo, S.H.; Bae, S.H.; Kim, E.J.; Joo, K.J.; Kuc, T.Y. Ontology-Based Knowledge Representation in Robotic Systems: A Survey Oriented toward Applications. Appl. Sci. 2021, 11, 4324. [Google Scholar] [CrossRef]
  59. Waibel, M.; Beetz, M.; Civera, J.; D’Andrea, R.; Elfring, J.; Gálvez-López, D.; Häussermann, K.; Janssen, R.; Montiel, J.; Perzylo, A.; et al. RoboEarth. IEEE Robot. Autom. Mag. 2011, 18, 69–82. [Google Scholar] [CrossRef]
  60. Daruna, A.; Liu, W.; Kira, Z.; Chetnova, S. RoboCSE: Robot Common Sense Embedding. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 9777–9783. [Google Scholar] [CrossRef]
  61. Ghallab, M.; Nau, D.S.; Traverso, P. Automated Planning—Theory and Practice; Morgan Kaufmann Publishers: San Francisco, CA, USA, 2004. [Google Scholar]
  62. Cashmore, M.; Fox, M.; Long, D.; Magazzeni, D.; Ridder, B.; Carrera, A.; Palomeras, N.; Hurtos, N.; Carreras, M. ROSPlan: Planning in the Robot Operating System. Proc. Int. Conf. Autom. Plan. Sched. 2015, 25, 333–341. [Google Scholar] [CrossRef]
  63. Rajan, K.; Py, F. T-REX: Partitioned inference for AUV mission control. In Proceedings of the Workshop on Planning and Plan Execution for Real-World Systems: Principles and Practices for Planning in Execution, Part of the The International Conference on Automated Planning and Scheduling (ICAPS 2007), Providence, RI, USA, 22–26 September 2007; pp. 171–199. [Google Scholar] [CrossRef]
  64. Narayanan, V. Human Factors Analysis of Automated Planning Technologies for Human-Robot. Ph.D. Thesis, Arizone State University, Tempe, AZ, USA, 2015. [Google Scholar]
  65. Raman, V.; Kress-Gazit, H. Explaining Impossible High-Level Robot Behaviors. IEEE Trans. Robot. 2013, 29, 94–104. [Google Scholar] [CrossRef]
  66. Madumal, P.; Miller, T.; Sonenberg, L.; Vetere, F. Explainable Reinforcement Learning through a Causal Lens. arXiv 2019, arXiv:1905.10958. [Google Scholar] [CrossRef]
  67. Stocking, K.C.; Gopnik, A.; Tomlin, C.J. From Robot Learning to Robot Understanding: Leveraging Causal Graphical Models for Robotics. In Proceedings of the 5th Conference on Robot Learning, London, UK, 8–11 November 2021; Volume 164, pp. 1776–1781. [Google Scholar]
  68. Brawer, J.; Qin, M.; Scassellati, B. A Causal Approach to Tool Affordance Learning. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24 January 2021; pp. 8394–8399. [Google Scholar] [CrossRef]
  69. Jara-Ettinger, J.; Gweon, H.; Schulz, L.E.; Tenenbaum, J.B. The Naïve Utility Calculus: Computational Principles Underlying Commonsense Psychology. Trends Cogn. Sci. 2016, 20, 589–604. [Google Scholar] [CrossRef] [PubMed]
  70. Huang, S.; Held, D.; Abbeel, P.; Dragan, A. Enabling Robots to Communicate their Objectives. Auton. Robot. 2019, 43, 309–326. [Google Scholar] [CrossRef]
  71. Zakershahrak, M.; Ghodratnama, S. Are We on The Same Page? Hierarchical Explanation Generation for Planning Tasks in Human-Robot Teaming Using Reinforcement Learning. arXiv 2020, arXiv:2012.11792. [Google Scholar]
  72. Hayes, B.; Shah, J.A. Improving Robot Controller Transparency Through Autonomous Policy Explanation. In Proceedings of the 2017 12th ACM/IEEE International Conference on Human-Robot Interaction HRI, Vienna, Austria, 6–9 March 2017; pp. 303–312. [Google Scholar]
  73. El Hafi, L.; Isobe, S.; Tabuchi, Y.; Katsumata, Y.; Nakamura, H.; Fukui, T.; Matsuo, T.; Garcia-Ricardez, G.; Yamamoto, M.; Taniguchi, A.; et al. System for augmented human–robot interaction through mixed reality and robot training by non-experts in customer service environments. Adv. Robot. 2020, 34, 157–172. [Google Scholar] [CrossRef]
  74. Romero-Garcés, A.; Martínez-Cruz, J.; Inglés-Romero, J.; Vicente-Chicote, C.; Marfil, R.; Bandera, A. Measuring Quality of Service in a Robotized Comprehensive Geriatric Assessment Scenario. Appl. Sci. 2020, 10, 6618. [Google Scholar] [CrossRef]
  75. Zakershahrak, M.; Marpally, S.R.; Sharma, A.; Gong, Z.; Zhang, Y. Order Matters: Generating Progressive Explanations for Planning Tasks in Human-Robot Teaming. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 3751–3757. [Google Scholar] [CrossRef]
  76. McDermott, D.; Ghallab, M.; Howe, A.E.; Knoblock, C.A.; Ram, A.; Veloso, M.M.; Weld, D.S.; Wilkins, D.E. PDDL-the Planning Domain Definition Language. Yale Center for Computational Vision and Control; Tech Report CVC TR-98.003/DCS TR-1165. 1998. Available online: https://courses.cs.washington.edu/courses/cse473/06sp/pddl.pdf (accessed on 16 March 2024).
  77. Haslum, P.; Lipovetzky, N.; Magazzeni, D.; Muise, C. An Introduction to the Planning Domain Definition Language. Synth. Lect. Artif. Intell. Mach. Learn. 2019, 13, 1–187. [Google Scholar] [CrossRef]
  78. Fox, M.; Long, D. PDDL2.1: An extension to PDDL for expressing temporal planning domains. J. Artif. Intell. Res. JAIR 2003, 20, 61–124. [Google Scholar] [CrossRef]
  79. Coradeschi, S.; Saffiotti, A. An introduction to the anchoring problem. Robot. Auton. Syst. 2003, 43, 85–96. [Google Scholar] [CrossRef]
  80. Bustos, P.; Manso, L.; Bandera, A.; Bandera, J.; García-Varea, I.; Martínez-Gómez, J. The CORTEX cognitive robotics architecture: Use cases. Cogn. Syst. Res. 2019, 55, 107–123. [Google Scholar] [CrossRef]
  81. Marfil, R.; Romero-Garcés, A.; Bandera, J.P.; Manso, L.J.; Calderita, L.V.; Bustos, P.; Bandera, A.; García-Polo, J.; Fernández, F.; Voilmy, D. Perceptions or Actions? Grounding How Agents Interact Within a Software Architecture for Cognitive Robotics. Cogn. Comput. 2020, 12, 479–497. [Google Scholar] [CrossRef]
  82. Mueller, S.T.; Hoffman, R.R.; Clancey, W.J.; Emrey, A.; Klein, G. Explanation in Human-AI Systems: A Literature Meta-Review, Synopsis of Key Ideas and Publications, and Bibliography for Explainable AI. arXiv 2019, arXiv:1902.01876. [Google Scholar]
  83. Smith, S.C.; Ramamoorthy, S. Counterfactual Explanation and Causal Inference In Service of Robustness in Robot Control. In Proceedings of the 2020 Joint IEEE 10th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob), Valparaiso, Chile, 26–30 October 2020; pp. 1–8. [Google Scholar] [CrossRef]
  84. Carreira-Perpiñán, M.; Hada, S. Counterfactual Explanations for Oblique Decision Trees:Exact, Efficient Algorithms. Proc. AAAI Conf. Artif. Intell. 2021, 35, 6903–6911. [Google Scholar] [CrossRef]
  85. Karimi, A.; Barthe, G.; Balle, B.; Valera, I. Model-Agnostic Counterfactual Explanations for Consequential Decisions. arXiv 2019, arXiv:1905.11190. [Google Scholar]
  86. Sokol, K.; Flach, P.A. Desiderata for Interpretability: Explaining Decision Tree Predictions with Counterfactuals. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019. [Google Scholar]
  87. Gjærum, V.B.; Strümke, I.; Løver, J.; Miller, T.; Lekkas, A.M. Model tree methods for explaining deep reinforcement learning agents in real-time robotic applications. Neurocomputing 2023, 515, 133–144. [Google Scholar] [CrossRef]
  88. Gjærum, V.B.; Strümke, I.; Alsos, O.A.; Lekkas, A.M. Explaining a Deep Reinforcement Learning Docking Agent Using Linear Model Trees with User Adapted Visualization. J. Mar. Sci. Eng. 2021, 9, 1178. [Google Scholar] [CrossRef]
  89. Hoffman, R.; Miller, T.; Mueller, S.T.; Klein, G.; Clancey, W.J. Explaining Explanation, Part 4: A Deep Dive on Deep Nets. IEEE Intell. Syst. 2018, 33, 87–95. [Google Scholar] [CrossRef]
  90. Trinidad Barnech, G.; Tejera, G.; Valle-Lisboa, J.; Núñez, P.; Bachiller, P.; Bustos, P. Initial Results with a Simulation Capable Robotics Cognitive Architecture. In ROBOT2022: Fifth Iberian Robotics Conference; Tardioli, D., Matellán, V., Heredia, G., Silva, M.F., Marques, L., Eds.; Springer International Publishing: Cham, Switzerland, 2023; pp. 612–623. [Google Scholar]
  91. Mania, P.; Kenfack, F.K.; Neumann, M.; Beetz, M. Imagination-enabled Robot Perception. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Prague, Czech Republic, 27 September–1 October 2021; pp. 936–943. [Google Scholar] [CrossRef]
  92. Kenghagho, F.K.; Neumann, M.; Mania, P.; Tan, T.; Siddiky, F.A.; Weller, R.; Zachmann, G.; Beetz, M. NaivPhys4RP—Towards Human-like Robot Perception ’Physical Reasoning based on Embodied Probabilistic Simulation’. In Proceedings of the IEEE-RAS International Conference on Humanoid Robots, Ginowan, Japan, 28–30 November 2022; pp. 815–822. [Google Scholar] [CrossRef]
  93. Bass, I.; Smith, K.; Bonawitz, E.; Ullman, T.D. Partial Mental Simulation Explains Fallacies in Physical Reasoning. Cogn. Neuropsychol. 2021, 38, 413–424. [Google Scholar] [CrossRef] [PubMed]
  94. Haidu, A.; Beßler, D.; Bozcuoğlu, A.K.; Beetz, M. KnowRobSIM—Game Engine-Enabled Knowledge Processing Towards Cognition-Enabled Robot Control. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 4491–4498. [Google Scholar] [CrossRef]
  95. Zhang, C.; Chen, J.; Li, J.; Peng, Y.; Mao, Z. Large language models for human–robot interaction: A review. Biomim. Intell. Robot. 2023, 3, 100131. [Google Scholar] [CrossRef]
  96. De Giacomo, G.; De Masellis, R.; Montali, M. Reasoning on LTL on finite traces: Insensitivity to infiniteness. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, AAAI’14, Québec City, QC, Canada, 27–31 July 2014; pp. 1027–1033. [Google Scholar]
  97. Gerevini, A.E.; Haslum, P.; Long, D.; Saetti, A.; Dimopoulos, Y. Deterministic planning in the fifth international planning competition: PDDL3 and experimental evaluation of the planners. Artif. Intell. 2009, 173, 619–668. [Google Scholar] [CrossRef]
  98. Bonassi, L.; Gerevini, A.E.; Scala, E. Planning with Qualitative Action-Trajectory Constraints in PDDL. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI-22, Vienna, Austria, 23–29 July 2022; Raedt, L.D., Ed.; pp. 4606–4613. [Google Scholar] [CrossRef]
  99. Jiang, Y.; Zhang, S.; Khandelwal, P.; Stone, P. Task planning in robotics: An empirical comparison of PDDL-and ASP-based systems. Front. Inf. Technol. Electron. Eng. 2019, 20, 363. [Google Scholar] [CrossRef]
Figure 1. A CLARA robot, equipped with a D435i camera from Intel, a pair of RFID antennas, and a face recognition-devoted hardware (Intel F455). The robot uses a SCITOS G3 as the base platform [34].
Figure 1. A CLARA robot, equipped with a D435i camera from Intel, a pair of RFID antennas, and a face recognition-devoted hardware (Intel F455). The robot uses a SCITOS G3 as the base platform [34].
Electronics 13 01169 g001
Figure 2. The three rungs in the Causality ladder (adapted from [35]).
Figure 2. The three rungs in the Causality ladder (adapted from [35]).
Electronics 13 01169 g002
Figure 3. The hierarchical mind model for a collaboration task represented by a spatial–temporal–causal And–Or graph [44]. The task “Picking up a trolley” implies that the robot must detect and approach the trolley. Once the robot is correctly positioned close to the trolley, it can either ask the person to manually place the trolley on its pallet truck (because, for example, space constraints prevent the robot from doing so) or it can move and pick it up itself.
Figure 3. The hierarchical mind model for a collaboration task represented by a spatial–temporal–causal And–Or graph [44]. The task “Picking up a trolley” implies that the robot must detect and approach the trolley. Once the robot is correctly positioned close to the trolley, it can either ask the person to manually place the trolley on its pallet truck (because, for example, space constraints prevent the robot from doing so) or it can move and pick it up itself.
Electronics 13 01169 g003
Figure 4. Schematic view of the counterfactual explanation generation in Gjærum et al. [14]’s proposal (see text for details).
Figure 4. Schematic view of the counterfactual explanation generation in Gjærum et al. [14]’s proposal (see text for details).
Electronics 13 01169 g004
Figure 5. Person and robot are internal plans π ¯ R R and π ¯ H H , respectively, generated with respect to their internal models. The person has no information about the robot’s model, or the robot is not explaining (or correctly explaining) its plan to the person. In any case, the problem is that the robot’s plan is interpreted by the person ( π R H ) with respect to his model Π H H [17].
Figure 5. Person and robot are internal plans π ¯ R R and π ¯ H H , respectively, generated with respect to their internal models. The person has no information about the robot’s model, or the robot is not explaining (or correctly explaining) its plan to the person. In any case, the problem is that the robot’s plan is interpreted by the person ( π R H ) with respect to his model Π H H [17].
Electronics 13 01169 g005
Figure 6. Perceptual anomaly and the search for a causal explanation.
Figure 6. Perceptual anomaly and the search for a causal explanation.
Electronics 13 01169 g006
Figure 7. KnowRobSYM software architecture (see Haidu et al. [94] for details).
Figure 7. KnowRobSYM software architecture (see Haidu et al. [94] for details).
Electronics 13 01169 g007
Figure 8. Number of papers including the topics Explanation and Robot. Web of Science Analyze filter: TS = (explanation) AND TS = (robot) (https://clarivate.com/webofsciencegroup/solutions/web-of-science/, accessed on 13 January 2024).
Figure 8. Number of papers including the topics Explanation and Robot. Web of Science Analyze filter: TS = (explanation) AND TS = (robot) (https://clarivate.com/webofsciencegroup/solutions/web-of-science/, accessed on 13 January 2024).
Electronics 13 01169 g008
Table 1. Knowledge representation in autonomous robotics and their relationships with explainability.
Table 1. Knowledge representation in autonomous robotics and their relationships with explainability.
ApproachKnowledge RepresentationStep (Figure 2)Application
Gao et al. [44]And–Or graphInterventionHuman–robot collaboration tasks
Han et al. [45]Behaviour TreesInterventionExplainable planning
Lindner and Olz [16]PDDL domain and E/D (causal) linksInterventionExplainable planning
Krarup et al. [46]Extended PDDL2.1CounterfactualsExplainable planning
Diehl and Ramirez-Amaro [47]Bayesian networksInterventionFailure explanation
Stange et al. [9]A collection of memories and modelsInterventionHuman–robot interaction
Alarcos et al. [48]Ontology for Collaborative Robotics and AdaptationInterventionHuman–robot collaboration tasks
Hanheide et al. [49]Three-layers hierarchical organizationCounterfactualsTask planning and execution in open and uncertain worlds. Failure explanation
Gjærum et al. [14]Linear Model TreesCounterfactualsObtaining counterfactuals explanations in real time for black-box models
Table 2. Issues for achieving an explainable robot [25].
Table 2. Issues for achieving an explainable robot [25].
RequirementSolutions
The sequential decision making Π R R must be interpretable by the person. That is, given A R R and τ R R
A R R : Π R R × τ R R
When π R R is derived using these and Π R R , without encountering an error, the decision-making space is defined as interpretable.
Π R R must be composed of states and actions natural to people, with transitions between states that are consistent with the real world. One solution is to manually encode the model using a collection of rules, but it is obviously complex to ensure that this set will be complete [25]. The solution can be to build a causal graph [47,66]. For instance, Stocking et al. [67] analyse how to distinguish task-relevant and -irrelevant variables exploring causality. Brawer et al. [68] propose an approach for a robot to learn an explicit model of cause-and-effect by constructing a structural causal model. Diehl and Ramirez-Amaro [47] describe an approach for learning a causal Bayesian network from simulated task executions. The obtained model was used for transferring the acquired knowledge from simulation to several real robots with different embodiments.
The robot must possess the tools ( Π H R , A H R , and  τ H R ) that allow it to infer the person’s behaviour.
A H R : Π H R × τ H R π H R
where π H R is a close estimate of π H H .
The Bayesian framework has been extensively used to predict human reasoning in goal-oriented tasks [69,70]. Gao et al. [44] propose to understand complex human activities using an action parsing algorithm based on a graph-based task representation. This graph allows the robot to infer human mental states in complex environments.
The robot must be able to generate explanations that allow the person to comprehend the robot’s plan π R R . This explanation should allow the person to have an estimate, as close as possible, to the plan executed by the robot, but not necessarily to understand the robot’s policy.Explanations must achieve a real interpretable decision-making space. In Figure 5, the situation has been exemplified in a very drastic way: the model that the person uses to interpret the robot’s plan is not influenced by the robot itself. Assuming that there is some influence (and that the model of the robot’s behaviour is estimated in some way by the person, Π R H ), and denoting by ϵ the explanation provided by the robot, the objective is that Π R H + ϵ allows the person to internalize a model Π ^ R H such that
A R H : Π ^ R H × τ R H π R R .
Different approaches have been proposed to generate explanations in the field of human–robot interaction. Typically, these works seek to provide explanations in a particular order or to interweave them with the execution of the plan. Since the person does not always want the same level of detail in the explanation, Zakershahrak and Ghodratnama [71] propose a hierarchical model of explanations and details that the robot requires to provide explanations as a target. Han et al. [45] propose to encode high-level robot explanations using Behaviour Trees (BTs), specifically giving answers to questions seeking causal information. In the proposal by Gao et al. [44], the internal model of the person is inferred and compared with the robot’s model. Whenever a disparity between these two models is detected, an explanation is generated to encourage a correction of the person’s internal state.
Explanations are not correctly encoded/conveyed to the personWe must note that solving Equation (11) is only the first step for providing an explanation. That is, after determining the explanation ϵ , the robot needs to encode it into a form that can be conveyed to people. This problem can be solved using verbalization [72] or visualization [73] schemes. However, in general, the goal should be to personalise the communication channel according to the preferences of the end users. In [74], the human–robot interaction employs several informative channels and this personalisation is achieved by using an adaptation framework. It is also relevant how the information is presented to the human. Most work on explanation generation assumes that the person understands an explanation, regardless of how much information it includes. As Zakershahrak et al. [26] pointed out, complex explanations should be provided in an online fashion, with each explanation broken down into multiple parts, which are then provided to the person separately and intertwined with the execution of the plan.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Marfil, R.; Bustos, P.; Bandera, A. Causal-Based Approaches to Explain and Learn from Self-Extension—A Review. Electronics 2024, 13, 1169. https://doi.org/10.3390/electronics13071169

AMA Style

Marfil R, Bustos P, Bandera A. Causal-Based Approaches to Explain and Learn from Self-Extension—A Review. Electronics. 2024; 13(7):1169. https://doi.org/10.3390/electronics13071169

Chicago/Turabian Style

Marfil, Rebeca, Pablo Bustos, and Antonio Bandera. 2024. "Causal-Based Approaches to Explain and Learn from Self-Extension—A Review" Electronics 13, no. 7: 1169. https://doi.org/10.3390/electronics13071169

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop