论文标题
网络压缩的部分正则化方法
A Partial Regularization Method for Network Compression
论文作者
论文摘要
深度神经网络取得了巨大的成功,依赖于随着网络深度和宽度的增加的GPU和大规模数据集的发展。但是,由于昂贵的计算和密集的记忆,研究人员专注于设计压缩方法,以便使它们适用于受限平台。在本文中,我们提出了一种部分正则化的方法,而不是惩罚所有参数的原始形式,据说是完全正则化的,以更高的速度进行模型压缩。根据神经网络的排列不变特性的存在,这是合理且可行的。实验结果表明,正如我们预期的那样,在几乎所有情况下观察到的运行时间较小,可以降低计算复杂性。应该是因为部分正则化方法会增强较低数量的计算元素。令人惊讶的是,它有助于改善一些重要的指标,例如回归拟合结果和在多个数据集上的培训和测试阶段的分类准确性,并告诉我们修剪模型具有更好的性能和泛化能力。更重要的是,我们分析结果并得出结论,即必须存在最佳网络结构并取决于输入数据。
Deep Neural Networks have achieved remarkable success relying on the developing availability of GPUs and large-scale datasets with increasing network depth and width. However, due to the expensive computation and intensive memory, researchers have concentrated on designing compression methods in order to make them practical for constrained platforms. In this paper, we propose an approach of partial regularization rather than the original form of penalizing all parameters, which is said to be full regularization, to conduct model compression at a higher speed. It is reasonable and feasible according to the existence of the permutation invariant property of neural networks. Experimental results show that as we expected, the computational complexity is reduced by observing less running time in almost all situations. It should be owing to the fact that partial regularization method invovles a lower number of elements for calculation. Surprisingly, it helps to improve some important metrics such as regression fitting results and classification accuracy in both training and test phases on multiple datasets, telling us that the pruned models have better performance and generalization ability. What's more, we analyze the results and draw a conclusion that an optimal network structure must exist and depend on the input data.