论文标题

通过最佳阈值修剪频道修剪

Channel Pruning via Optimal Thresholding

论文作者

Ye, Yun, You, Ganmei, Fwu, Jong-Kae, Zhu, Xia, Yang, Qing, Zhu, Yuan

论文摘要

结构化的修剪,尤其是通道修剪被广泛用于降低的计算成本以及与现成的硬件设备的兼容性。在现有作品中,通常使用预定义的全局阈值或根据预定义的度量计算的阈值来删除权重。基于预定义的全球阈值设计忽略了不同层和权重分布之间的变化,因此,它们通常会导致过度闭塞或欠不足引起的次优性能。在本文中,我们提出了一种简单而有效的方法,称为最佳阈值(OT),以缩减具有层依赖性阈值的修剪通道,这些阈值与可忽略不计的通道最佳地分开。通过使用OT,大多数可忽略不计或不重要的通道可以修剪以实现高稀疏性,同时最大程度地减少性能降解。由于保留了最重要的权重,因此可以进一步调整修剪模型并迅速收敛,很少迭代。我们的方法表明了卓越的性能,尤其是与高稀疏性的最新设计相比。在CIFAR-100上,使用OT仅使用1.46E8拖鞋和0.71亿参数实现了75.99%的精度,并通过使用OT实现75.99%的精度。

Structured pruning, especially channel pruning is widely used for the reduced computational cost and the compatibility with off-the-shelf hardware devices. Among existing works, weights are typically removed using a predefined global threshold, or a threshold computed from a predefined metric. The predefined global threshold based designs ignore the variation among different layers and weights distribution, therefore, they may often result in sub-optimal performance caused by over-pruning or under-pruning. In this paper, we present a simple yet effective method, termed Optimal Thresholding (OT), to prune channels with layer dependent thresholds that optimally separate important from negligible channels. By using OT, most negligible or unimportant channels are pruned to achieve high sparsity while minimizing performance degradation. Since most important weights are preserved, the pruned model can be further fine-tuned and quickly converge with very few iterations. Our method demonstrates superior performance, especially when compared to the state-of-the-art designs at high levels of sparsity. On CIFAR-100, a pruned and fine-tuned DenseNet-121 by using OT achieves 75.99% accuracy with only 1.46e8 FLOPs and 0.71M parameters.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源