论文标题

脱钩的侧信息融合以进行顺序推荐

Decoupled Side Information Fusion for Sequential Recommendation

论文作者

Xie, Yueqi, Zhou, Peilin, Kim, Sunghun

论文摘要

顺序推荐(SR)的侧信息融合旨在有效利用各种侧面信息来增强下一项目预测的性能。大多数最先进的方法都建立在自我发项网络的基础上,并专注于探索各种解决方案,以在注意力层之前集成项目嵌入和侧面信息嵌入。但是,我们的分析表明,各种类型嵌入的早期整合限制了由于等级瓶颈而引起的注意矩阵的表现力,并限制了梯度的灵活性。同样,它涉及不同异质信息资源之间的混合相关性,这给注意力计算带来了额外的干扰。在此激励的情况下,我们提出了隔离的侧面信息融合以进行顺序推荐(DIF-SR),该信息将侧面信息从输入移至注意力层,并将各种侧面信息和项目表示的注意力计算分解。从理论上讲,我们从理论上和经验上表明,所提出的解决方案允许更高的注意矩阵和柔性梯度增强侧信息融合的建模能力。同样,提出了辅助属性预测因子,以进一步激活侧面信息和项目表示学习之间的有益相互作用。在四个现实世界数据集上进行的广泛实验表明,我们提出的解决方案稳定地胜过最先进的SR模型。进一步的研究表明,我们提出的解决方案可以很容易地纳入当前的基于注意力的SR模型中,并显着提高了性能。我们的源代码可在https://github.com/aim-se/dif-sr上找到。

Side information fusion for sequential recommendation (SR) aims to effectively leverage various side information to enhance the performance of next-item prediction. Most state-of-the-art methods build on self-attention networks and focus on exploring various solutions to integrate the item embedding and side information embeddings before the attention layer. However, our analysis shows that the early integration of various types of embeddings limits the expressiveness of attention matrices due to a rank bottleneck and constrains the flexibility of gradients. Also, it involves mixed correlations among the different heterogeneous information resources, which brings extra disturbance to attention calculation. Motivated by this, we propose Decoupled Side Information Fusion for Sequential Recommendation (DIF-SR), which moves the side information from the input to the attention layer and decouples the attention calculation of various side information and item representation. We theoretically and empirically show that the proposed solution allows higher-rank attention matrices and flexible gradients to enhance the modeling capacity of side information fusion. Also, auxiliary attribute predictors are proposed to further activate the beneficial interaction between side information and item representation learning. Extensive experiments on four real-world datasets demonstrate that our proposed solution stably outperforms state-of-the-art SR models. Further studies show that our proposed solution can be readily incorporated into current attention-based SR models and significantly boost performance. Our source code is available at https://github.com/AIM-SE/DIF-SR.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源