论文标题
感知可视化:透过DNN的眼睛看
Perception Visualization: Seeing Through the Eyes of a DNN
论文作者
论文摘要
人工智能(AI)系统能力我们所生活的世界。深层神经网络(DNNS)能够在场景的不断扩展的情况下解决任务,但是我们渴望采用这些强大的模型的渴望使我们专注于他们的性能并剥夺我们理解它们的能力。当前在可解释的AI领域的研究试图通过开发各种扰动或基于梯度的解释技术来弥合这一差距。对于图像,这些技术无法完全捕获并传达所需的语义信息,以阐明模型为何做出预测。在这项工作中,我们开发了一种新的解释形式,该解释本质上与当前的解释方法(例如Grad-CAM)完全不同。感知可视化通过描述潜在表示与对应的视觉模式相对应的视觉图像,提供了DNN在输入图像中感知的内容的视觉表示。可视化是通过反转编码特征的重建模型获得的,以使原始模型的参数和预测未经修改。我们的用户研究的结果表明,当人们可以使用感知可视化时,人类可以更好地理解和预测系统的决策,从而减轻深层模型作为可信系统的调试和部署。
Artificial intelligence (AI) systems power the world we live in. Deep neural networks (DNNs) are able to solve tasks in an ever-expanding landscape of scenarios, but our eagerness to apply these powerful models leads us to focus on their performance and deprioritises our ability to understand them. Current research in the field of explainable AI tries to bridge this gap by developing various perturbation or gradient-based explanation techniques. For images, these techniques fail to fully capture and convey the semantic information needed to elucidate why the model makes the predictions it does. In this work, we develop a new form of explanation that is radically different in nature from current explanation methods, such as Grad-CAM. Perception visualization provides a visual representation of what the DNN perceives in the input image by depicting what visual patterns the latent representation corresponds to. Visualizations are obtained through a reconstruction model that inverts the encoded features, such that the parameters and predictions of the original models are not modified. Results of our user study demonstrate that humans can better understand and predict the system's decisions when perception visualizations are available, thus easing the debugging and deployment of deep models as trusted systems.