论文标题
神经网络中的等距表示改善鲁棒性
Isometric Representations in Neural Networks Improve Robustness
论文作者
论文摘要
人工和生物学代理商会在完全随机和非结构化的数据中学习。数据结构在数据点之间的度量关系中编码。在神经网络的背景下,层中的神经元活动形成一个表示该层在其输入上实现的转换的表示。为了以真实的方式利用数据中的结构,这种表示应反映输入距离,因此是连续和等距的。在支持这一陈述的神经科学中的最新发现表明,概括和鲁棒性与神经表示不断相关。在机器学习中,大多数算法缺乏鲁棒性,通常认为依赖于与人类使用的数据不同的各个方面,正如对抗性攻击中常见的那样。在跨凝结分类期间,网络表示的度量和结构特性通常在类之间和内部都被打破。训练的这种副作用会导致在无法保留这种结构的位置附近的扰动下不稳定。获得鲁棒性的标准解决方案之一是添加临时正规化项,但据我们所知,迫使表示形式保留输入数据的度量结构作为稳定机制。在这项工作中,我们训练神经网络以执行分类,同时维持课堂内的度量结构,从而导致等距内表示。这种网络表示形式有益于准确和鲁棒的推断。通过使用此属性堆叠层,我们创建了一个网络体系结构,以促进内部神经表示的层次操纵。最后,我们验证等距正则化改善了对MNIST的对抗性攻击的鲁棒性。
Artificial and biological agents cannon learn given completely random and unstructured data. The structure of data is encoded in the metric relationships between data points. In the context of neural networks, neuronal activity within a layer forms a representation reflecting the transformation that the layer implements on its inputs. In order to utilize the structure in the data in a truthful manner, such representations should reflect the input distances and thus be continuous and isometric. Supporting this statement, recent findings in neuroscience propose that generalization and robustness are tied to neural representations being continuously differentiable. In machine learning, most algorithms lack robustness and are generally thought to rely on aspects of the data that differ from those that humans use, as is commonly seen in adversarial attacks. During cross-entropy classification, the metric and structural properties of network representations are usually broken both between and within classes. This side effect from training can lead to instabilities under perturbations near locations where such structure is not preserved. One of the standard solutions to obtain robustness is to add ad hoc regularization terms, but to our knowledge, forcing representations to preserve the metric structure of the input data as a stabilising mechanism has not yet been studied. In this work, we train neural networks to perform classification while simultaneously maintaining within-class metric structure, leading to isometric within-class representations. Such network representations turn out to be beneficial for accurate and robust inference. By stacking layers with this property we create a network architecture that facilitates hierarchical manipulation of internal neural representations. Finally, we verify that isometric regularization improves the robustness to adversarial attacks on MNIST.