论文标题

联合矩阵分解:算法设计和应用到数据群集

Federated Matrix Factorization: Algorithm Design and Application to Data Clustering

论文作者

Wang, Shuai, Chang, Tsung-Hui

论文摘要

对数据隐私的最新需求要求将联合学习(FL)作为大规模和异质网络中的新分布式学习范式。尽管已经提出了许多FL算法,但其中很少有人考虑矩阵分解(MF)模型,该模型已知具有大量的信号处理和机器学习应用。与现有的FL算法不同,该算法是针对联合MF(FEDMF)在单个变量中平稳问题的,必须处理具有挑战性的非凸和非平滑问题(由于约束或正则化),并使用两个变量块。在本文中,我们通过提出两种新的FEDMF算法,即基于模型平均和梯度共享原则来应对挑战。 FEDMAVG和FEDMGS都采用了每个通信的多个局部更新的步骤,以加快融合的速度,并且仅允许随机采样的客户端集与服务器通信以降低通信成本。分别介绍了两种算法的收敛分析,其中描述了数据分布,本地更新号码和部分客户端通信对算法性能的影响。通过关注数据聚类任务,提出了广泛的实验结果,以检查算法的实际性能,并证明了它们对现有分布式聚类算法的疗效。

Recent demands on data privacy have called for federated learning (FL) as a new distributed learning paradigm in massive and heterogeneous networks. Although many FL algorithms have been proposed, few of them have considered the matrix factorization (MF) model, which is known to have a vast number of signal processing and machine learning applications. Different from the existing FL algorithms that are designed for smooth problems with single block of variables, in federated MF (FedMF), one has to deal with challenging non-convex and non-smooth problems (due to constraints or regularization) with two blocks of variables. In this paper, we address the challenge by proposing two new FedMF algorithms, namely, FedMAvg and FedMGS, based on the model averaging and gradient sharing principles, respectively. Both FedMAvg and FedMGS adopt multiple steps of local updates per communication round to speed up convergence, and allow only a randomly sampled subset of clients to communicate with the server for reducing the communication cost. Convergence analyses for the two algorithms are respectively presented, which delineate the impacts of data distribution, local update number, and partial client communication on the algorithm performance. By focusing on a data clustering task, extensive experiment results are presented to examine the practical performance of both algorithms, as well as demonstrating their efficacy over the existing distributed clustering algorithms.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源