论文标题
点击序列:通过良好的固定建筑修剪,有效且准确的端到端深度学习培训
ClickTrain: Efficient and Accurate End-to-End Deep Learning Training via Fine-Grained Architecture-Preserving Pruning
论文作者
论文摘要
卷积神经网络(CNN)由于对预测准确性和分析质量的需求的增长而变得越来越深,更广泛和非线性。但是,宽大的CNN需要大量的计算资源和处理时间。许多以前的作品都研究了修剪以提高推理性能的模型,但是几乎没有做出有效降低培训成本的工作。在本文中,我们提出了点击序列:有效,准确的端到端培训和CNN的修剪框架。与现有的修剪训练工作不同,ClickTrain通过细粒度固定的修剪提供了更高的模型准确性和压缩比。通过利用基于模式的修剪,我们提出的新型精确的重量重要性估计,动态模式产生和选择以及编译器辅助计算的优化,ClickTrain与基线训练相比,CLICKTrain生成了高度准确且快速的CNN CNN模型,用于直接部署而无需任何额外的时间间隔。 ClickTrain还以可比的精度和压缩比,将修剪后训练方法的端到端时间成本降低了2.3倍。此外,与最先进的修剪训练方法相比,ClickTrain在相似的有限培训时间下,在经过测试的CNN模型和数据集上提供了显着的精度和压缩率。
Convolutional neural networks (CNNs) are becoming increasingly deeper, wider, and non-linear because of the growing demand on prediction accuracy and analysis quality. The wide and deep CNNs, however, require a large amount of computing resources and processing time. Many previous works have studied model pruning to improve inference performance, but little work has been done for effectively reducing training cost. In this paper, we propose ClickTrain: an efficient and accurate end-to-end training and pruning framework for CNNs. Different from the existing pruning-during-training work, ClickTrain provides higher model accuracy and compression ratio via fine-grained architecture-preserving pruning. By leveraging pattern-based pruning with our proposed novel accurate weight importance estimation, dynamic pattern generation and selection, and compiler-assisted computation optimizations, ClickTrain generates highly accurate and fast pruned CNN models for direct deployment without any extra time overhead, compared with the baseline training. ClickTrain also reduces the end-to-end time cost of the pruning-after-training method by up to 2.3X with comparable accuracy and compression ratio. Moreover, compared with the state-of-the-art pruning-during-training approach, ClickTrain provides significant improvements both accuracy and compression ratio on the tested CNN models and datasets, under similar limited training time.