论文标题

用噪声激励块增强卷积神经网络

Enhance Convolutional Neural Networks with Noise Incentive Block

论文作者

Xia, Menghan, Wang, Yi, Han, Chu, Wong, Tien-Tsin

论文摘要

作为一种通用的建模工具,卷积神经网络(CNN)已被广泛用于图像生成和翻译任务。但是,当用平面输入馈送时,由于空间共享的卷积内核,当前的CNN模型可能无法产生生动的结果。我们称其为CNN的平坦降解。不幸的是,这种降解是从平面输入中产生空间变化的输出的最大障碍,在先前的文献中几乎没有讨论过。为了解决这个问题,我们提出了一个模型不可知的解决方案,即噪声激励块(NIB),该解决方案是任何CNN生成模型的通用插件。关键的想法是打破平坦的输入条件,同时保持原始信息的完整性。具体而言,NIB与噪声图对称地散布输入数据,并将其重新组合在由目标函数驱动的特征域中。广泛的实验表明,配备有NIB的现有CNN模型从平坦的降解中得以生存,并且能够在某些特定图像生成任务中产生更丰富的详细信息,以产生更丰富的详细信息,例如。语义图像综合,数据隐藏图像产生和深层神经抖动。

As a generic modeling tool, Convolutional Neural Networks (CNNs) have been widely employed in image generation and translation tasks. However, when fed with a flat input, current CNN models may fail to generate vivid results due to the spatially shared convolution kernels. We call it the flatness degradation of CNNs. Unfortunately, such degradation is the greatest obstacles to generate a spatially-variant output from a flat input, which has been barely discussed in the previous literature. To tackle this problem, we propose a model agnostic solution, i.e. Noise Incentive Block (NIB), which serves as a generic plug-in for any CNN generation model. The key idea is to break the flat input condition while keeping the intactness of the original information. Specifically, the NIB perturbs the input data symmetrically with a noise map and reassembles them in the feature domain as driven by the objective function. Extensive experiments show that existing CNN models equipped with NIB survive from the flatness degradation and are able to generate visually better results with richer details in some specific image generation tasks given flat inputs, e.g. semantic image synthesis, data-hidden image generation, and deep neural dithering.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源