论文标题
发现图形神经网络的不变理由
Discovering Invariant Rationales for Graph Neural Networks
论文作者
论文摘要
图神经网络(GNN)的固有解释性是找到输入图特征的一小部分 - 理由 - 指导模型预测。不幸的是,领先的合理化模型通常依赖于数据偏见,尤其是快捷特征来构成理由并做出预测,而无需探究关键和因果模式。此外,这些数据偏见很容易在培训分布之外发生变化。结果,这些模型在分发数据上的可解释性和预测性能下降。在这项工作中,我们提出了一种新的策略,以发现不变理由(DIR)来构建本质上可解释的GNN。它对培训分配进行干预,以创建多个介入分配。然后,它接近在不同分布中不变的因果原理,同时滤除不稳定的虚假模式。在合成数据集和现实世界数据集的实验验证了DIR的优越性,其对图分类的可解释性和泛化能力而不是领先的基线。代码和数据集可从https://github.com/wuyxin/dir-gnn获得。
Intrinsic interpretability of graph neural networks (GNNs) is to find a small subset of the input graph's features -- rationale -- which guides the model prediction. Unfortunately, the leading rationalization models often rely on data biases, especially shortcut features, to compose rationales and make predictions without probing the critical and causal patterns. Moreover, such data biases easily change outside the training distribution. As a result, these models suffer from a huge drop in interpretability and predictive performance on out-of-distribution data. In this work, we propose a new strategy of discovering invariant rationale (DIR) to construct intrinsically interpretable GNNs. It conducts interventions on the training distribution to create multiple interventional distributions. Then it approaches the causal rationales that are invariant across different distributions while filtering out the spurious patterns that are unstable. Experiments on both synthetic and real-world datasets validate the superiority of our DIR in terms of interpretability and generalization ability on graph classification over the leading baselines. Code and datasets are available at https://github.com/Wuyxin/DIR-GNN.