论文标题

自我监督发现可解释的增强功能

Self-Supervised Discovering of Interpretable Features for Reinforcement Learning

论文作者

Shi, Wenjie, Huang, Gao, Song, Shiji, Wang, Zhuoyuan, Lin, Tingyu, Wu, Cheng

论文摘要

深度强化学习(RL)最近导致了一系列复杂的控制任务的许多突破。但是,代理商的决策过程通常不透明。缺乏解释性阻碍了RL在安全 - 关键方案中的适用性。尽管几种方法试图解释基于视觉的RL,但大多数方法都没有对代理行为的详细解释。在本文中,我们提出了一个自制的可解释框架,该框架可以发现可解释的特征,以便于对非专家的RL代理轻松理解。具体而言,采用自我监督的可解释网络(SSINET)来产生精细的注意力面具,以突出与任务相关的信息,这构成了代理商决策的大多数证据。我们在几次Atari 2600游戏以及Duckietown上验证和评估我们的方法,这是一个充满挑战的自动驾驶汽车模拟器环境。结果表明,我们的方法提供了有关代理如何做出决策以及代理商表现出色或出色的经验证据,尤其是在转移到新型场景时。总体而言,我们的方法为基于视觉的RL的内部决策过程提供了宝贵的见解。此外,我们的方法不使用任何外部标记的数据,因此证明了通过自我监督的方式学习高质量面具的可能性,这可能会阐明新的范式,以用于无标记的视觉学习,例如自我监督的细分和检测。

Deep reinforcement learning (RL) has recently led to many breakthroughs on a range of complex control tasks. However, the agent's decision-making process is generally not transparent. The lack of interpretability hinders the applicability of RL in safety-critical scenarios. While several methods have attempted to interpret vision-based RL, most come without detailed explanation for the agent's behavior. In this paper, we propose a self-supervised interpretable framework, which can discover interpretable features to enable easy understanding of RL agents even for non-experts. Specifically, a self-supervised interpretable network (SSINet) is employed to produce fine-grained attention masks for highlighting task-relevant information, which constitutes most evidence for the agent's decisions. We verify and evaluate our method on several Atari 2600 games as well as Duckietown, which is a challenging self-driving car simulator environment. The results show that our method renders empirical evidences about how the agent makes decisions and why the agent performs well or badly, especially when transferred to novel scenes. Overall, our method provides valuable insight into the internal decision-making process of vision-based RL. In addition, our method does not use any external labelled data, and thus demonstrates the possibility to learn high-quality mask through a self-supervised manner, which may shed light on new paradigms for label-free vision learning such as self-supervised segmentation and detection.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源