论文标题
与自学的修剪卷积神经网络
Pruning Convolutional Neural Networks with Self-Supervision
论文作者
论文摘要
在没有监督的情况下训练的卷积神经网络几乎与受监督的预训练相匹配,但有时是以更高数量的参数为代价。从保留性能的这些大型无监督的弯头中提取子网特别是使它们减少计算密集型的感兴趣。典型的修剪方法在培训任务期间操作,同时试图维持同一任务上修剪的网络的性能。但是,在自我监管的特征学习中,培训目标是关于表示下游任务的表示性转移性的不可知论。因此,保留此目标的性能并不能确保修剪的子网在解决下游任务方面仍然有效。在这项工作中,我们研究了主要用于监督学习的标准修剪方法的使用,用于未经标签的网络(即,在自我监督任务上)。我们表明,在标签上重新训练时,带有或没有标签的修剪面膜可以达到可比性的性能,这表明修剪在自我监督和监督的学习中相似。有趣的是,我们还发现,修剪保留了自我监督的子网表示的转移绩效。
Convolutional neural networks trained without supervision come close to matching performance with supervised pre-training, but sometimes at the cost of an even higher number of parameters. Extracting subnetworks from these large unsupervised convnets with preserved performance is of particular interest to make them less computationally intensive. Typical pruning methods operate during training on a task while trying to maintain the performance of the pruned network on the same task. However, in self-supervised feature learning, the training objective is agnostic on the representation transferability to downstream tasks. Thus, preserving performance for this objective does not ensure that the pruned subnetwork remains effective for solving downstream tasks. In this work, we investigate the use of standard pruning methods, developed primarily for supervised learning, for networks trained without labels (i.e. on self-supervised tasks). We show that pruned masks obtained with or without labels reach comparable performance when re-trained on labels, suggesting that pruning operates similarly for self-supervised and supervised learning. Interestingly, we also find that pruning preserves the transfer performance of self-supervised subnetwork representations.