论文标题

大规模项目建议的自我监督学习

Self-supervised Learning for Large-scale Item Recommendations

论文作者

Yao, Tiansheng, Yi, Xinyang, Cheng, Derek Zhiyuan, Yu, Felix, Chen, Ting, Menon, Aditya, Hong, Lichan, Chi, Ed H., Tjoa, Steve, Kang, Jieqi, Ettinger, Evan

论文摘要

大型推荐模型从大量目录中找到大多数相关项目,它们在现代搜索和推荐系统中起着至关重要的作用。为了建模具有大磁带分类特征的输入空间,典型的推荐模型通过神经网络从用户反馈数据中学习了一个通过神经网络的联合嵌入空间。但是,由于语料库中的数百万到数十亿到数十亿美元,用户倾向于为其中一小部分提供反馈,从而导致幂律分布。这使得长尾项目的反馈数据极为稀疏。 受到最新在计算机视觉和自然语言理解方面的自我监督代表性学习研究的成功启发,我们为大规模项目建议提出了一个多任务自我监督学习(SSL)框架。该框架旨在通过学习项目特征的更好的潜在关系来解决标签稀疏问题。具体而言,SSL改善了项目表示学习,并用作额外的正则化以改善概括。此外,我们提出了一种新型的数据增强方法,该方法利用了所提出的框架内的特征相关性。 我们使用两个现实世界数据集评估了我们的框架,分别具有500m和1B培训示例。我们的结果证明了SSL正则化的有效性,并显示了其优于最新正则化技术的表现。我们还已经向Web规模的商业应用程序到应用程序推荐系统推出了拟议的技术,在A/B实验实验中展示了有关实时流量的顶级业务指标的重大改进。我们的在线业绩还验证了我们的假设,即我们的框架确实在缺乏监督的切片上改善了模型性能。

Large scale recommender models find most relevant items from huge catalogs, and they play a critical role in modern search and recommendation systems. To model the input space with large-vocab categorical features, a typical recommender model learns a joint embedding space through neural networks for both queries and items from user feedback data. However, with millions to billions of items in the corpus, users tend to provide feedback for a very small set of them, causing a power-law distribution. This makes the feedback data for long-tail items extremely sparse. Inspired by the recent success in self-supervised representation learning research in both computer vision and natural language understanding, we propose a multi-task self-supervised learning (SSL) framework for large-scale item recommendations. The framework is designed to tackle the label sparsity problem by learning better latent relationship of item features. Specifically, SSL improves item representation learning as well as serving as additional regularization to improve generalization. Furthermore, we propose a novel data augmentation method that utilizes feature correlations within the proposed framework. We evaluate our framework using two real-world datasets with 500M and 1B training examples respectively. Our results demonstrate the effectiveness of SSL regularization and show its superior performance over the state-of-the-art regularization techniques. We also have already launched the proposed techniques to a web-scale commercial app-to-app recommendation system, with significant improvements top-tier business metrics demonstrated in A/B experiments on live traffic. Our online results also verify our hypothesis that our framework indeed improves model performance even more on slices that lack supervision.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源