Publikationen (FIS)

Explainable Reinforcement Learning via Dynamic Mixture Policies

authored by
Maximilian Schier, Frederik Schubert, Bodo Rosenhahn
Abstract

Learning control policies using deep reinforcement learning has shown great success for a variety of applications, including robotics and automated driving. A key area limiting the adaptation of RL in the real world is the lack of trust in the decision-making process of such policies. Therefore, explainability is a requirement of any RL agent operating in the real world. In this work, we propose a family of control policies that are explainable-by-design regarding individual observation components on object-based scene representations. By estimating diagonal squashed Gaussian and categorical mixture distributions on sub-spaces of the decomposed observations, we develop stochastic policies with easy-to-read explanations of the decision-making process. Our design is generally applicable to any RL algorithm using stochastic policies. We showcase the explainability on an extensive suite of single-and multi-agent simulations, set-and sequence-based high-level scenes, and discrete and continuous action spaces, with performance at least on-par or better compared to standard policy architectures. In additional experiments, we analyze the robustness of our approach to its single additional hyper-parameter and examine its potential for very low computational requirements with tiny policies.

Organisation(s)
L3S Research Centre
Institute of Information Processing
Type
Conference contribution
Pages
13530-13536
No. of pages
7
Publication date
19.05.2025
Publication status
Published
Peer reviewed
Yes
ASJC Scopus subject areas
Software, Control and Systems Engineering, Artificial Intelligence, Electrical and Electronic Engineering
Electronic version(s)
https://doi.org/10.1109/ICRA55743.2025.11127782 (Access: Closed)