论文标题
解释性作为统计推断
Explainability as statistical inference
论文作者
论文摘要
近年来,已经提出了各种各样的模型解释方法,所有这些方法都以不同的理由和启发式为导向。在本文中,我们采用了一条新的路线,并将可解释性作为统计推断问题。我们提出了一个旨在产生可解释预测的一般深层概率模型。可以通过最大可能性来学习模型参数,并且该方法可以适应任何预测器网络体系结构和任何类型的预测问题。我们的方法是一种摊销的可解释性模型的情况,其中神经网络被用作选择器,以便在推理时间进行快速解释。几种流行的可解释性方法被证明是我们一般模型的正则最大可能性的特殊情况。我们提出了具有地面真相选择的新数据集,以评估功能重要性图。使用这些数据集,我们通过实验表明,使用多个插补提供了更合理的解释。
A wide variety of model explanation approaches have been proposed in recent years, all guided by very different rationales and heuristics. In this paper, we take a new route and cast interpretability as a statistical inference problem. We propose a general deep probabilistic model designed to produce interpretable predictions. The model parameters can be learned via maximum likelihood, and the method can be adapted to any predictor network architecture and any type of prediction problem. Our method is a case of amortized interpretability models, where a neural network is used as a selector to allow for fast interpretation at inference time. Several popular interpretability methods are shown to be particular cases of regularised maximum likelihood for our general model. We propose new datasets with ground truth selection which allow for the evaluation of the features importance map. Using these datasets, we show experimentally that using multiple imputation provides more reasonable interpretations.