论文标题
八卦并参加:上下文敏感的图表表示学习
Gossip and Attend: Context-Sensitive Graph Representation Learning
论文作者
论文摘要
图表示学习(GRL)是一种强大的技术,用于学习高维和稀疏图的低维矢量表示。大多数研究都使用随机步行探索与图相关的结构和元数据,并采用无监督或半监督的学习方案。在这些方法中学习是不含上下文的,每个节点只有一个表示。最近的研究提出了单个表示和提出的上下文敏感方法的充分性,这些方法能够为不同上下文提取多个节点表示。事实证明,这在诸如链接预测和排名之类的应用中非常有效。 但是,这些方法中的大多数都依赖于需要复杂且昂贵的RNN或CNN来捕获高级功能或依靠社区检测算法来识别节点的多个上下文的其他文本功能。 在这项研究中,我们表明,要提取高质量上下文敏感的节点表示形式,不需要依靠补充节点特征,也不需要采用计算沉重且复杂的模型。我们提出了山羊,这是一种受八卦通信启发的上下文敏感算法,并且仅在图表结构上是一种相互关注的机制。我们在链接预测和节点群集任务上使用6个现实世界数据集显示了山羊的功效,并将其与12个流行和最新的(SOTA)基线进行比较。山羊一贯胜过它们,并且在链接预测和聚类任务上分别获得了最佳性能方法,最多可获得12%和19%的增长。
Graph representation learning (GRL) is a powerful technique for learning low-dimensional vector representation of high-dimensional and often sparse graphs. Most studies explore the structure and metadata associated with the graph using random walks and employ an unsupervised or semi-supervised learning schemes. Learning in these methods is context-free, resulting in only a single representation per node. Recently studies have argued on the adequacy of a single representation and proposed context-sensitive approaches, which are capable of extracting multiple node representations for different contexts. This proved to be highly effective in applications such as link prediction and ranking. However, most of these methods rely on additional textual features that require complex and expensive RNNs or CNNs to capture high-level features or rely on a community detection algorithm to identify multiple contexts of a node. In this study we show that in-order to extract high-quality context-sensitive node representations it is not needed to rely on supplementary node features, nor to employ computationally heavy and complex models. We propose GOAT, a context-sensitive algorithm inspired by gossip communication and a mutual attention mechanism simply over the structure of the graph. We show the efficacy of GOAT using 6 real-world datasets on link prediction and node clustering tasks and compare it against 12 popular and state-of-the-art (SOTA) baselines. GOAT consistently outperforms them and achieves up to 12% and 19% gain over the best performing methods on link prediction and clustering tasks, respectively.