论文标题
迈向基于原型的自我解释的图形神经网络
Towards Prototype-Based Self-Explainable Graph Neural Network
论文作者
论文摘要
图神经网络(GNN)在为各种域建模图形结构数据方面表现出很高的能力。但是,GNN被称为缺乏可解释性的黑盒模型。在不了解他们的内在工作的情况下,我们无法完全信任他们,这在很大程度上限制了他们在高风险场景中的采用。尽管已经采取了一些初步的努力来解释GNN的预测,但它们主要集中于使用其他解释器提供事后解释,这可能会歪曲目标GNN的真正内部工作机制。关于自我解释的GNN的作品相当有限。因此,我们研究了学习基于原型的自我解释的新问题,可以同时提供对预测的准确预测和基于原型的解释。我们设计了一个可以学习原型图的框架,以捕获每个类的代表性模式作为班级的解释。学到的原型还用于同时对测试实例进行预测,并提供实例级别的解释。对现实世界和合成数据集的广泛实验显示了提出的框架在预测准确性和解释质量方面的有效性。
Graph Neural Networks (GNNs) have shown great ability in modeling graph-structured data for various domains. However, GNNs are known as black-box models that lack interpretability. Without understanding their inner working, we cannot fully trust them, which largely limits their adoption in high-stake scenarios. Though some initial efforts have been taken to interpret the predictions of GNNs, they mainly focus on providing post-hoc explanations using an additional explainer, which could misrepresent the true inner working mechanism of the target GNN. The works on self-explainable GNNs are rather limited. Therefore, we study a novel problem of learning prototype-based self-explainable GNNs that can simultaneously give accurate predictions and prototype-based explanations on predictions. We design a framework which can learn prototype graphs that capture representative patterns of each class as class-level explanations. The learned prototypes are also used to simultaneously make prediction for for a test instance and provide instance-level explanation. Extensive experiments on real-world and synthetic datasets show the effectiveness of the proposed framework for both prediction accuracy and explanation quality.