论文标题
ProtoryNet-可解释的文本分类通过原型轨迹
ProtoryNet - Interpretable Text Classification Via Prototype Trajectories
论文作者
论文摘要
我们根据新的原型轨迹概念提出了一个新颖的可解释的深层神经网络,用于文本分类,称为ProtoryNet。 ProtoryNet在现代语言学中的原型理论中的动机,通过在文本序列中找到每个句子的最相似原型,并以每个句子的距离与相应的有效原型的距离喂食RNN主链,从而进行预测。然后,RNN主链捕获原型的时间模式,我们称之为原型轨迹。原型轨迹能够对RNN模型的推理过程进行直观和细粒度的解释,与人类如何分析文本相似。我们还设计了一个原型修剪程序,以减少模型使用的原型总数,以更好地解释性。多个公共数据集的实验表明,ProtoryNet比基线原型的深神经网络更准确,并且与最新的黑盒模型相比,ProtoryNet可减少性能差距。此外,在原型修剪后,所得的ProtoryNet模型仅需要小于或大约20个原型的所有数据集,这显着有益于可解释性。此外,我们报告了一个调查结果,表明人类用户发现ProtoryNet比其他基于原型的方法更直观,更易于理解。
We propose a novel interpretable deep neural network for text classification, called ProtoryNet, based on a new concept of prototype trajectories. Motivated by the prototype theory in modern linguistics, ProtoryNet makes a prediction by finding the most similar prototype for each sentence in a text sequence and feeding an RNN backbone with the proximity of each sentence to the corresponding active prototype. The RNN backbone then captures the temporal pattern of the prototypes, which we refer to as prototype trajectories. Prototype trajectories enable intuitive and fine-grained interpretation of the reasoning process of the RNN model, in resemblance to how humans analyze texts. We also design a prototype pruning procedure to reduce the total number of prototypes used by the model for better interpretability. Experiments on multiple public data sets show that ProtoryNet is more accurate than the baseline prototype-based deep neural net and reduces the performance gap compared to state-of-the-art black-box models. In addition, after prototype pruning, the resulting ProtoryNet models only need less than or around 20 prototypes for all datasets, which significantly benefits interpretability. Furthermore, we report a survey result indicating that human users find ProtoryNet more intuitive and easier to understand than other prototype-based methods.