Score: 1

Temporal Counterfactual Explanations of Behaviour Tree Decisions

Published: September 9, 2025 | arXiv ID: 2509.07674v1

By: Tamlin Love, Antonio Andriella, Guillem Alenyà

Potential Business Impact:

Helps robots explain why they do things.

Business Areas:
Robotics Hardware, Science and Engineering, Software

Explainability is a critical tool in helping stakeholders understand robots. In particular, the ability for robots to explain why they have made a particular decision or behaved in a certain way is useful in this regard. Behaviour trees are a popular framework for controlling the decision-making of robots and other software systems, and thus a natural question to ask is whether or not a system driven by a behaviour tree is capable of answering "why" questions. While explainability for behaviour trees has seen some prior attention, no existing methods are capable of generating causal, counterfactual explanations which detail the reasons for robot decisions and behaviour. Therefore, in this work, we introduce a novel approach which automatically generates counterfactual explanations in response to contrastive "why" questions. Our method achieves this by first automatically building a causal model from the structure of the behaviour tree as well as domain knowledge about the state and individual behaviour tree nodes. The resultant causal model is then queried and searched to find a set of diverse counterfactual explanations. We demonstrate that our approach is able to correctly explain the behaviour of a wide range of behaviour tree structures and states. By being able to answer a wide range of causal queries, our approach represents a step towards more transparent, understandable and ultimately trustworthy robotic systems.

Country of Origin
🇪🇸 Spain

Repos / Data Links

Page Count
23 pages

Category
Computer Science:
Robotics