Score: 2

A Survey of Explainable Reinforcement Learning: Targets, Methods and Needs

Published: July 16, 2025 | arXiv ID: 2507.12599v1

By: Léo Saulières

Potential Business Impact:

Explains how smart computer programs make decisions.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

The success of recent Artificial Intelligence (AI) models has been accompanied by the opacity of their internal mechanisms, due notably to the use of deep neural networks. In order to understand these internal mechanisms and explain the output of these AI models, a set of methods have been proposed, grouped under the domain of eXplainable AI (XAI). This paper focuses on a sub-domain of XAI, called eXplainable Reinforcement Learning (XRL), which aims to explain the actions of an agent that has learned by reinforcement learning. We propose an intuitive taxonomy based on two questions "What" and "How". The first question focuses on the target that the method explains, while the second relates to the way the explanation is provided. We use this taxonomy to provide a state-of-the-art review of over 250 papers. In addition, we present a set of domains close to XRL, which we believe should get attention from the community. Finally, we identify some needs for the field of XRL.

Repos / Data Links

Page Count
69 pages

Category
Computer Science:
Artificial Intelligence