论文标题

NERN-学习神经网络的神经表示

NeRN -- Learning Neural Representations for Neural Networks

论文作者

Ashkenazi, Maor, Rimon, Zohar, Vainshtein, Ron, Levi, Shir, Richardson, Elad, Mintz, Pinchas, Treister, Eran

论文摘要

最近已显示神经表示可以有效地重建从3D网格和形状到图像和视频的广泛信号。我们表明,正确调整后,可以使用神经表示直接表示预训练的卷积神经网络的权重,从而导致神经网络(NERN)的神经表示。受到先前神经表示方法的坐标输入的启发,我们根据其在架构中的位置为网络中的每个卷积内核分配了一个坐标,并优化了预测器网络以将坐标映射到其相应的权重。与视觉场景的空间平滑度相似,我们表明,在原始网络的重量上纳入平滑度的约束有助于Nern对更好的重建。此外,由于预训练的模型权重的轻微扰动可能会导致相当精确的损失,因此我们采用了知识蒸馏领域的技术来稳定学习过程。我们证明了NERN在CIFAR-10,CIFAR-100和Imagenet上重建广泛使用的体系结构中的有效性。最后,我们使用NERN提出了两个应用程序,证明了学习表示的能力。

Neural Representations have recently been shown to effectively reconstruct a wide range of signals from 3D meshes and shapes to images and videos. We show that, when adapted correctly, neural representations can be used to directly represent the weights of a pre-trained convolutional neural network, resulting in a Neural Representation for Neural Networks (NeRN). Inspired by coordinate inputs of previous neural representation methods, we assign a coordinate to each convolutional kernel in our network based on its position in the architecture, and optimize a predictor network to map coordinates to their corresponding weights. Similarly to the spatial smoothness of visual scenes, we show that incorporating a smoothness constraint over the original network's weights aids NeRN towards a better reconstruction. In addition, since slight perturbations in pre-trained model weights can result in a considerable accuracy loss, we employ techniques from the field of knowledge distillation to stabilize the learning process. We demonstrate the effectiveness of NeRN in reconstructing widely used architectures on CIFAR-10, CIFAR-100, and ImageNet. Finally, we present two applications using NeRN, demonstrating the capabilities of the learned representations.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源