论文标题

神经网络很难学习生活游戏

It's Hard for Neural Networks To Learn the Game of Life

论文作者

Springer, Jacob M., Kenyon, Garrett T.

论文摘要

提高神经网络学习能力的努力主要集中在优化方法的作用上,而不是重量初始化。然而,最近的发现表明,神经网络依赖于称为“彩票”的子网的幸运随机初始权重,这些重量迅速融合到解决方案中。为了调查重量初始化如何影响性能,我们检查了经过训练的小型卷积网络,以预测二维蜂窝蜂窝自动机康威的生命游戏的n个步骤,可以在2n+1层卷积网络中有效实现其更新规则。我们发现,该任务训练的该架构的网络很少收敛。相反,网络需要更多的参数才能始终如一地收敛。此外,近乎最小的体系结构对参数的微小变化很敏感:更改单个重量的符号可能会导致网络无法学习。最后,我们观察到一个临界值D_0,以便训练最小网络,其中具有示例的示例,其中细胞充满活力,概率D_0显着增加了收敛的机会。我们得出的结论是,培训卷积神经网络学习以生命游戏的n个步骤表示的输入/输出功能表现出了彩票票证假设所预测的许多特征,即学习此功能所需的网络的大小通常大于实施该功能所需的最小网络。

Efforts to improve the learning abilities of neural networks have focused mostly on the role of optimization methods rather than on weight initializations. Recent findings, however, suggest that neural networks rely on lucky random initial weights of subnetworks called "lottery tickets" that converge quickly to a solution. To investigate how weight initializations affect performance, we examine small convolutional networks that are trained to predict n steps of the two-dimensional cellular automaton Conway's Game of Life, the update rules of which can be implemented efficiently in a 2n+1 layer convolutional network. We find that networks of this architecture trained on this task rarely converge. Rather, networks require substantially more parameters to consistently converge. In addition, near-minimal architectures are sensitive to tiny changes in parameters: changing the sign of a single weight can cause the network to fail to learn. Finally, we observe a critical value d_0 such that training minimal networks with examples in which cells are alive with probability d_0 dramatically increases the chance of convergence to a solution. We conclude that training convolutional neural networks to learn the input/output function represented by n steps of Game of Life exhibits many characteristics predicted by the lottery ticket hypothesis, namely, that the size of the networks required to learn this function are often significantly larger than the minimal network required to implement the function.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源