论文标题
张量分解以压缩深度学习中的卷积层
Tensor decomposition to Compress Convolutional Layers in Deep Learning
论文作者
论文摘要
张量数据的特征提取是许多任务中的重要步骤,例如异常检测,过程监视,图像分类和质量控制。尽管已经提出了许多用于张量特征提取的方法,但仍有两个挑战需要解决:1)如何降低高维和大量张量数据的计算成本; 2)如何解释输出功能并评估其重要性。 {深度学习中的最新方法,例如卷积神经网络(CNN),在分析张量数据方面表现出了出色的性能,但是模型的复杂性和缺乏可解释性仍然阻碍了它们的广泛采用。为了填补这一研究差距,我们建议使用CP分解来在深度学习中近似压缩卷积层(CPAC-CONV层)。我们的工作的贡献可以概括为三个方面:(1)我们适应了CP分类以压缩卷积内核,并为我们提出的CPAC-CONV层提供了向前和向后传播的表达; (2)与原始卷积层相比,提出的CPAC-CONV层可以减少参数的数量而不会衰减预测性能。它可以与其他层结合起来建立新的深神经网络。 (3)分解内核的值表示相应特征图的重要性,该特征图为我们提供了指导特征选择的见解。
Feature extraction for tensor data serves as an important step in many tasks such as anomaly detection, process monitoring, image classification, and quality control. Although many methods have been proposed for tensor feature extraction, there are still two challenges that need to be addressed: 1) how to reduce the computation cost for high dimensional and large volume tensor data; 2) how to interpret the output features and evaluate their significance. {The most recent methods in deep learning, such as Convolutional Neural Network (CNN), have shown outstanding performance in analyzing tensor data, but their wide adoption is still hindered by model complexity and lack of interpretability. To fill this research gap, we propose to use CP-decomposition to approximately compress the convolutional layer (CPAC-Conv layer) in deep learning. The contributions of our work could be summarized into three aspects: (1) we adapt CP-decomposition to compress convolutional kernels and derive the expressions of both forward and backward propagations for our proposed CPAC-Conv layer; (2) compared with the original convolutional layer, the proposed CPAC-Conv layer can reduce the number of parameters without decaying prediction performance. It can combine with other layers to build novel deep Neural Networks; (3) the value of decomposed kernels indicates the significance of the corresponding feature map, which provides us with insights to guide feature selection.