论文标题

通过基于进化的搜索进行对Mobilenet进行二进制Mobilenet

Binarizing MobileNet via Evolution-based Searching

论文作者

Phan, Hai, Liu, Zechun, Huynh, Dang, Savvides, Marios, Cheng, Kwang-Ting, Shen, Zhiqiang

论文摘要

二进制神经网络(BNN),已知是有效的紧凑网络体系结构之一,在视觉任务中取得了巨大的结果。由于网络的二进制性质,设计有效的二进制架构并不是微不足道的。在本文中,我们建议使用进化搜索,以促进构造和培训方案,当时将Mobilenet(一种具有可分开的深度卷积的紧凑网络进行二手化。受一声体系结构搜索框架的启发,我们操纵了团体卷积的想法,以设计有效的1位卷积神经网络(CNN),假设计算成本和模型准确性之间的折衷大约是最佳的权衡。我们的目标是通过探索小组卷积的最佳候选者,同时以复杂性和潜伏期来优化模型性能,从而提出一个微小而有效的二进制神经架构。该方法是三倍。首先,我们在每个卷积层都具有广泛的随机组组合训练强基线二进制网络。该设置使二进制神经网络具有通过层次保存基本信息的能力。其次,为了找到一组良好的超参数用于小组卷积,我们利用了利用有效1位模型探索的进化搜索。最后,这些二进制模型以通常的方式从头开始训练,以实现最终的二进制模型。进行了成像网的各种实验,以表明遵循我们的施工指南,最终模型可实现60.09%的TOP-1准确性,并以相同的计算成本胜过最先进的CI-BCNN。

Binary Neural Networks (BNNs), known to be one among the effectively compact network architectures, have achieved great outcomes in the visual tasks. Designing efficient binary architectures is not trivial due to the binary nature of the network. In this paper, we propose a use of evolutionary search to facilitate the construction and training scheme when binarizing MobileNet, a compact network with separable depth-wise convolution. Inspired by one-shot architecture search frameworks, we manipulate the idea of group convolution to design efficient 1-Bit Convolutional Neural Networks (CNNs), assuming an approximately optimal trade-off between computational cost and model accuracy. Our objective is to come up with a tiny yet efficient binary neural architecture by exploring the best candidates of the group convolution while optimizing the model performance in terms of complexity and latency. The approach is threefold. First, we train strong baseline binary networks with a wide range of random group combinations at each convolutional layer. This set-up gives the binary neural networks a capability of preserving essential information through layers. Second, to find a good set of hyperparameters for group convolutions we make use of the evolutionary search which leverages the exploration of efficient 1-bit models. Lastly, these binary models are trained from scratch in a usual manner to achieve the final binary model. Various experiments on ImageNet are conducted to show that following our construction guideline, the final model achieves 60.09% Top-1 accuracy and outperforms the state-of-the-art CI-BCNN with the same computational cost.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源