论文标题
基于梯度的竞争学习:理论
Gradient-based Competitive Learning: Theory
论文作者
论文摘要
深度学习已被广泛用于监督学习和分类/回归问题。最近,一个新的研究领域将这种范式应用于无监督的任务。实际上,一种基于梯度的方法有效,自主地提取了处理输入数据的相关功能。但是,最新技术主要集中在算法效率和准确性上,而不是模仿输入歧管。相反,竞争性学习是复制输入分布拓扑的强大工具。本文通过结合这两种技术来介绍该领域的新观点:基于梯度和竞争性学习。该理论是基于直觉,即神经网络能够直接在输入矩阵的转换上来学习拓扑结构。为此,介绍了香草竞争层及其双重层。前者只是标准竞争层的深度聚类的改编,而后者则是在转移矩阵上训练的。它们的等效性在理论上和实验上都得到了广泛的证明。但是,双层更适合处理非常高维数据集。提出的方法具有很大的潜力,因为它可以推广到许多拓扑学习任务,例如非平稳和等级聚类。此外,它也可以集成到更复杂的体系结构中,例如自动编码器和生成对抗网络。
Deep learning has been widely used for supervised learning and classification/regression problems. Recently, a novel area of research has applied this paradigm to unsupervised tasks; indeed, a gradient-based approach extracts, efficiently and autonomously, the relevant features for handling input data. However, state-of-the-art techniques focus mostly on algorithmic efficiency and accuracy rather than mimic the input manifold. On the contrary, competitive learning is a powerful tool for replicating the input distribution topology. This paper introduces a novel perspective in this area by combining these two techniques: unsupervised gradient-based and competitive learning. The theory is based on the intuition that neural networks are able to learn topological structures by working directly on the transpose of the input matrix. At this purpose, the vanilla competitive layer and its dual are presented. The former is just an adaptation of a standard competitive layer for deep clustering, while the latter is trained on the transposed matrix. Their equivalence is extensively proven both theoretically and experimentally. However, the dual layer is better suited for handling very high-dimensional datasets. The proposed approach has a great potential as it can be generalized to a vast selection of topological learning tasks, such as non-stationary and hierarchical clustering; furthermore, it can also be integrated within more complex architectures such as autoencoders and generative adversarial networks.