TY - GEN
T1 - Explain to Gain
T2 - 3rd World Conference on Explainable Artificial Intelligence, xAI 2025
AU - Quintana-Amate, Santiago
AU - Stevens, Delaney
AU - Ketheeswaran, Varniethan
AU - Capaldo, Patrick
AU - Sheldon, Dylan
AU - Hall, Mark
PY - 2009/7/31
Y1 - 2009/7/31
N2 - This article presents a new method for leveraging explainable reinforcement learning (XRL) knowledge to enhance the performance of reinforcement learning (RL) agents. Although current XRL approaches are mainly focused on improving interpretability and user trust by providing explanations for agent actions, their ability to guide and optimise RL agent’s training is under-explored. To address this gap, we extend an existing introspective analysis framework by integrating XRL metrics directly into the training pipelines of model-free RL algorithms. This integration allows dynamic adjustments of algorithm-specific parameters based on real-time feedback from XRL metrics. The proposed methodology is validated across diverse OpenAI Gym environments (CartPole and Taxi). By evaluating both on-policy and off-policy approaches, we demonstrate that incorporating XRL insights leads to significant improvements in agent performance. The analysis of the results highlights the benefits regarding enhanced explainability and optimised decision-making. This work contributes in XRL research area by aligning interpretability with actionable performance gains, paving the way for more reliable and transparent RL systems in complex, real-world applications.
AB - This article presents a new method for leveraging explainable reinforcement learning (XRL) knowledge to enhance the performance of reinforcement learning (RL) agents. Although current XRL approaches are mainly focused on improving interpretability and user trust by providing explanations for agent actions, their ability to guide and optimise RL agent’s training is under-explored. To address this gap, we extend an existing introspective analysis framework by integrating XRL metrics directly into the training pipelines of model-free RL algorithms. This integration allows dynamic adjustments of algorithm-specific parameters based on real-time feedback from XRL metrics. The proposed methodology is validated across diverse OpenAI Gym environments (CartPole and Taxi). By evaluating both on-policy and off-policy approaches, we demonstrate that incorporating XRL insights leads to significant improvements in agent performance. The analysis of the results highlights the benefits regarding enhanced explainability and optimised decision-making. This work contributes in XRL research area by aligning interpretability with actionable performance gains, paving the way for more reliable and transparent RL systems in complex, real-world applications.
KW - Dynamic Algorithm Adjustment
KW - Explainability Metrics
KW - Explainable autonomous agents
KW - Explainable Reinforcement Learning (XRL)
KW - Exploration-Exploitation Trade-off
KW - Introspective Reinforcement Learning (IxDRL)
UR - https://www.scopus.com/pages/publications/105020675532
U2 - 10.1007/978-3-032-08324-1_11
DO - 10.1007/978-3-032-08324-1_11
M3 - Chapter in a published conference proceeding
AN - SCOPUS:105020675532
SN - 9783032083234
T3 - Communications in Computer and Information Science
SP - 247
EP - 270
BT - Explainable Artificial Intelligence - 3rd World Conference, xAI 2025, Proceedings
A2 - Guidotti, Riccardo
A2 - Schmid, Ute
A2 - Longo, Luca
PB - Springer
CY - Cham, Switzerland
Y2 - 9 July 2025 through 11 July 2025
ER -