论文标题
探索CNN的参数可重复使用性
Exploring the parameter reusability of CNN
论文作者
论文摘要
最近,使用小型数据训练网络已成为深度学习领域的热门话题。重复培训的参数是解决半监督和转移学习问题的最重要策略之一。但是,这些方法成功的基本原因尚不清楚。在本文中,我们提出了一种解决方案,该解决方案不仅可以根据重复使用卷积内核的性能进行判断是否可以重复使用,而且还可以根据重复使用相应参数的性能,最终判断哪些层的特定网络参数可以重复使用,最终,这些参数是基于均值的均值(rmse se sse sse sse os n overs necorse of Square)的目标。具体而言,我们定义了CNN参数重用的成功取决于两个条件:首先,网络是可重复使用的网络;其次,来自源域和目标域的卷积内核之间的RMSE足够小。实验结果表明,在满足这些条件时,将重复使用参数的性能显着改善。
In recent times, using small data to train networks has become a hot topic in the field of deep learning. Reusing pre-trained parameters is one of the most important strategies to address the issue of semi-supervised and transfer learning. However, the fundamental reason for the success of these methods is still unclear. In this paper, we propose a solution that can not only judge whether a given network is reusable or not based on the performance of reusing convolution kernels but also judge which layers' parameters of the given network can be reused, based on the performance of reusing corresponding parameters and, ultimately, judge whether those parameters are reusable or not in a target task based on the root mean square error (RMSE) of the corresponding convolution kernels. Specifically, we define that the success of a CNN's parameter reuse depends upon two conditions: first, the network is a reusable network; and second, the RMSE between the convolution kernels from the source domain and target domain is small enough. The experimental results demonstrate that the performance of reused parameters applied to target tasks, when these conditions are met, is significantly improved.