论文标题

深网的平稳激活和可重复性

Smooth activations and reproducibility in deep networks

论文作者

Shamir, Gil I., Lin, Dong, Coviello, Lorenzo

论文摘要

由于它们的惊人成功,深层网络几乎逐渐穿透了我们生活中的每个领域。但是,具有实质性的性能准确性提高的价格是\ emph {不复制性}的价格。在完全相同的培训数据集上训练的两个相同的模型,即使平均精度相似,在单个示例上的预测上也可能显示出很大的差异,尤其是在对高度分布的平行系统进行培训时。流行的整流线性单元(RELU)激活一直是深层网络最近成功的关键。但是,我们证明了Relu也是深层网络中不可培养的催化剂。我们表明,不仅可以比Relu更顺畅地提供更好的准确性,而且还可以提供更好的准确性 - 可重复使用权的权衡。我们提出了一个新的激活家庭。平滑的relu(\ emph {smelu}),旨在给出这样更好的权衡,同时还保持数学表达式简单,从而使实施便宜。 Smelu是单调的,模仿的Relu,同时提供连续的梯度,可产生更好的可重复性。我们将SMELU推广到具有更大的灵活性,然后证明Smelu及其广义形式是更通用的平滑连续单元(RESCU)激活的更通用方法的特殊情况。经验结果表明,尤其是SMELU的光滑激活,具有卓越的准确性可得到的权衡。

Deep networks are gradually penetrating almost every domain in our lives due to their amazing success. However, with substantive performance accuracy improvements comes the price of \emph{irreproducibility}. Two identical models, trained on the exact same training dataset may exhibit large differences in predictions on individual examples even when average accuracy is similar, especially when trained on highly distributed parallel systems. The popular Rectified Linear Unit (ReLU) activation has been key to recent success of deep networks. We demonstrate, however, that ReLU is also a catalyzer to irreproducibility in deep networks. We show that not only can activations smoother than ReLU provide better accuracy, but they can also provide better accuracy-reproducibility tradeoffs. We propose a new family of activations; Smooth ReLU (\emph{SmeLU}), designed to give such better tradeoffs, while also keeping the mathematical expression simple, and thus implementation cheap. SmeLU is monotonic, mimics ReLU, while providing continuous gradients, yielding better reproducibility. We generalize SmeLU to give even more flexibility and then demonstrate that SmeLU and its generalized form are special cases of a more general methodology of REctified Smooth Continuous Unit (RESCU) activations. Empirical results demonstrate the superior accuracy-reproducibility tradeoffs with smooth activations, SmeLU in particular.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源