论文标题

自动点:自动调整卷积神经网络,以改善转移学习

AutoTune: Automatically Tuning Convolutional Neural Networks for Improved Transfer Learning

论文作者

Basha, S. H. Shabbeer, Vinakota, Sravan Kumar, Pulabaigari, Viswanath, Mukherjee, Snehasis, Dubey, Shiv Ram

论文摘要

通过使用大规模数据集训练的预训练的深层网络,转移学习可以解决具有有限数据的特定任务。通常,在将学习的知识从源任务转移到目标任务时,最后几层在目标数据集上进行了微调(重新训练)。但是,这些层最初是为可能不适合目标任务的源任务而设计的。在本文中,我们介绍了一种自动调整卷积神经网络(CNN)以改善转移学习的机制。预先训练的CNN层使用贝叶斯优化的目标数据调整了目标数据的知识。首先,我们通过替换SoftMax层中神经元的数量的数量,用目标任务中的类数来训练基本CNN模型的最后一层。接下来,通过在验证数据(贪婪标准)上观察分类性能来自动调整预训练的CNN。为了评估所提出方法的性能,在三个基准数据集上进行了实验,例如Caltech-101,Caltech-256和Stanford Dogs。通过提议的自动点方法获得的分类结果优于三个数据集的标准基线传输方法,分别达到了$ 95.92 \%$,$ 86.54 \%\%\%$ $,而$ 84.67 \%$ $ $ $ $ $ $ $ $精度分别是Caltech-101,Caltech-101,Caltech-256和Stanford Dogs,以及Stanford Dogs。在这项研究中获得的实验结果表明,使用目标数据集的知识对预训练的CNN层进行调整供应更好的转移学习能力。源代码可在https://github.com/jekyllandhyde8999/autotune_cnn_transferlearning获得。

Transfer learning enables solving a specific task having limited data by using the pre-trained deep networks trained on large-scale datasets. Typically, while transferring the learned knowledge from source task to the target task, the last few layers are fine-tuned (re-trained) over the target dataset. However, these layers are originally designed for the source task that might not be suitable for the target task. In this paper, we introduce a mechanism for automatically tuning the Convolutional Neural Networks (CNN) for improved transfer learning. The pre-trained CNN layers are tuned with the knowledge from target data using Bayesian Optimization. First, we train the final layer of the base CNN model by replacing the number of neurons in the softmax layer with the number of classes involved in the target task. Next, the pre-trained CNN is tuned automatically by observing the classification performance on the validation data (greedy criteria). To evaluate the performance of the proposed method, experiments are conducted on three benchmark datasets, e.g., CalTech-101, CalTech-256, and Stanford Dogs. The classification results obtained through the proposed AutoTune method outperforms the standard baseline transfer learning methods over the three datasets by achieving $95.92\%$, $86.54\%$, and $84.67\%$ accuracy over CalTech-101, CalTech-256, and Stanford Dogs, respectively. The experimental results obtained in this study depict that tuning of the pre-trained CNN layers with the knowledge from the target dataset confesses better transfer learning ability. The source codes are available at https://github.com/JekyllAndHyde8999/AutoTune_CNN_TransferLearning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源