论文标题
解决通用图卷积网络的过度光滑
Tackling Over-Smoothing for General Graph Convolutional Networks
论文作者
论文摘要
提高GCN的深度(预计将允许更多表达性)显示出损害性能的损害,尤其是在节点分类上。原因的主要原因在于过度平滑。过度平滑的问题将GCN的输出驱动到一个在节点之间包含有限的区别信息的空间,从而导致表现不佳。已经提出了一些有关完善GCN架构的作品,但从理论上讲,这些改进是否能够缓解过度平滑。在本文中,我们首先从理论上分析了通用GCN的深度增加,包括通用GCN,GCN,具有偏见,RESGCN和APPNP。我们发现所有这些模型都以通用过程为特征:所有节点都会融合到立方体。在这个定理上,我们提议通过在每个训练时期随机去除一定数量的边缘,以减轻过度平滑。从理论上讲,Dropedge要么降低过度平滑的收敛速度,要么减轻了尺寸崩溃引起的信息损失。对模拟数据集的实验评估已经可视化了不同GCN之间过度平滑的差异。此外,对几个真正的基准支持的广泛实验,这些实验始终提高了各种浅GCN和深度GCN的性能。
Increasing the depth of GCN, which is expected to permit more expressivity, is shown to incur performance detriment especially on node classification. The main cause of this lies in over-smoothing. The over-smoothing issue drives the output of GCN towards a space that contains limited distinguished information among nodes, leading to poor expressivity. Several works on refining the architecture of deep GCN have been proposed, but it is still unknown in theory whether or not these refinements are able to relieve over-smoothing. In this paper, we first theoretically analyze how general GCNs act with the increase in depth, including generic GCN, GCN with bias, ResGCN, and APPNP. We find that all these models are characterized by a universal process: all nodes converging to a cuboid. Upon this theorem, we propose DropEdge to alleviate over-smoothing by randomly removing a certain number of edges at each training epoch. Theoretically, DropEdge either reduces the convergence speed of over-smoothing or relieves the information loss caused by dimension collapse. Experimental evaluations on simulated dataset have visualized the difference in over-smoothing between different GCNs. Moreover, extensive experiments on several real benchmarks support that DropEdge consistently improves the performance on a variety of both shallow and deep GCNs.