论文标题
Hcore-Init:基于图脱落的神经网络初始化
Hcore-Init: Neural Network Initialization based on Graph Degeneracy
论文作者
论文摘要
神经网络是人工智能的巅峰之作,因为近年来我们目睹了许多新颖的体系结构,学习和优化技术的深度学习。利用这一事实是,神经网络固有地构成神经元之间的多部分图,我们旨在直接分析它们的结构,以提取可以改善学习过程的有意义的信息。对于我们的知识图挖掘技术,尚未对神经网络中的学习进行增强。在本文中,我们为从深度学习体系结构中提取的完整加权多部分图的K核结构提出了一个改编版本。由于多片图是两分图的组合,而两部分图的组合是超图的起点图,因此我们设计了k-hypercore分解,这是k核退化性的超图类似物。我们将K-Hypercore应用于几个神经网络架构,更具体地用于卷积神经网络和多层感知,以进行图像识别任务,后者进行了非常短的训练。然后,我们使用神经元的超核数量提供的信息来重新定位神经网络的权重,从而偏向梯度优化方案。广泛的实验证明,K-Hypercore的表现优于最新初始化方法。
Neural networks are the pinnacle of Artificial Intelligence, as in recent years we witnessed many novel architectures, learning and optimization techniques for deep learning. Capitalizing on the fact that neural networks inherently constitute multipartite graphs among neuron layers, we aim to analyze directly their structure to extract meaningful information that can improve the learning process. To our knowledge graph mining techniques for enhancing learning in neural networks have not been thoroughly investigated. In this paper we propose an adapted version of the k-core structure for the complete weighted multipartite graph extracted from a deep learning architecture. As a multipartite graph is a combination of bipartite graphs, that are in turn the incidence graphs of hypergraphs, we design k-hypercore decomposition, the hypergraph analogue of k-core degeneracy. We applied k-hypercore to several neural network architectures, more specifically to convolutional neural networks and multilayer perceptrons for image recognition tasks after a very short pretraining. Then we used the information provided by the hypercore numbers of the neurons to re-initialize the weights of the neural network, thus biasing the gradient optimization scheme. Extensive experiments proved that k-hypercore outperforms the state-of-the-art initialization methods.