论文标题

培训事项:解锁更深的图形卷积神经网络的潜力

Training Matters: Unlocking Potentials of Deeper Graph Convolutional Neural Networks

论文作者

Luan, Sitao, Zhao, Mingde, Chang, Xiao-Wen, Precup, Doina

论文摘要

图形卷积网络(GCN)的性能限制以及我们不能堆叠更多的事实来提高性能,通常我们通常会为其他深度学习范式所做的绩效,这是由于GCN层的限制而引起的,这是由于GCN层的限制而引起的,包括不足的表达能力等。但是,如果如此,我们只能通过固定的培训来降低训练的训练,从而不再能够改善该步骤,并且无法改善该步骤,并且无法改善该步骤,并且无法改善该步骤,并且无法改善该步骤,并且无法改善该步骤,并且不再能够改善该过程。可能但可能以几种方式。本文首先从图形信号能量损失的角度确定GCN的训练难度。更具体地说,我们发现在训练期间向后通过的能量损失无效,使对输入更接近的层的学习无效。然后,我们提出了几种方法,以从能量的角度稍微修改GCN操作员来减轻训练问题。经过经验验证后,我们确认操作员的这些变化导致训练困难和显着性能的显着减少,而无需更改参数的组成。有了这些,我们得出的结论是,问题的根本原因比其他问题更可能是训练难度。

The performance limit of Graph Convolutional Networks (GCNs) and the fact that we cannot stack more of them to increase the performance, which we usually do for other deep learning paradigms, are pervasively thought to be caused by the limitations of the GCN layers, including insufficient expressive power, etc. However, if so, for a fixed architecture, it would be unlikely to lower the training difficulty and to improve performance by changing only the training procedure, which we show in this paper not only possible but possible in several ways. This paper first identify the training difficulty of GCNs from the perspective of graph signal energy loss. More specifically, we find that the loss of energy in the backward pass during training nullifies the learning of the layers closer to the input. Then, we propose several methodologies to mitigate the training problem by slightly modifying the GCN operator, from the energy perspective. After empirical validation, we confirm that these changes of operator lead to significant decrease in the training difficulties and notable performance boost, without changing the composition of parameters. With these, we conclude that the root cause of the problem is more likely the training difficulty than the others.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源