论文标题
同时:注意机制与多时间嵌入的混合物进行顺序建议
MEANTIME: Mixture of Attention Mechanisms with Multi-temporal Embeddings for Sequential Recommendation
论文作者
论文摘要
最近,基于自我注意力的模型在顺序推荐任务中实现了最先进的绩效。遵循语言处理的习俗,这些模型中的大多数都依赖于简单的位置嵌入来利用用户历史记录的顺序性质。但是,关于当前方法存在一些局限性。首先,顺序建议与语言处理不同,因为可以使用时间戳信息。以前的模型尚未充分利用它来提取其他上下文信息。其次,使用简单的嵌入方案可以导致信息瓶颈,因为相同的嵌入必须代表所有可能的上下文偏见。第三,由于以前的模型在每个注意力头上都使用相同的位置嵌入,因此它们可以浪费地学习重叠的模式。为了解决这些局限性,我们建议使用多种类型的时间嵌入,旨在捕获用户行为序列中的各种模式,以及一种充分利用如此多样性的注意力结构,该同时(将注意力机制与多阶段嵌入式的混合物混合在一起)。关于现实世界数据的实验表明,我们提出的方法的表现优于当前最新的顺序建议方法,我们提供了广泛的消融研究,以分析该模型如何从各种位置信息中获取。
Recently, self-attention based models have achieved state-of-the-art performance in sequential recommendation task. Following the custom from language processing, most of these models rely on a simple positional embedding to exploit the sequential nature of the user's history. However, there are some limitations regarding the current approaches. First, sequential recommendation is different from language processing in that timestamp information is available. Previous models have not made good use of it to extract additional contextual information. Second, using a simple embedding scheme can lead to information bottleneck since the same embedding has to represent all possible contextual biases. Third, since previous models use the same positional embedding in each attention head, they can wastefully learn overlapping patterns. To address these limitations, we propose MEANTIME (MixturE of AtteNTIon mechanisms with Multi-temporal Embeddings) which employs multiple types of temporal embeddings designed to capture various patterns from the user's behavior sequence, and an attention structure that fully leverages such diversity. Experiments on real-world data show that our proposed method outperforms current state-of-the-art sequential recommendation methods, and we provide an extensive ablation study to analyze how the model gains from the diverse positional information.