论文标题
用于联合学习的强大数量意识聚合
Robust Quantity-Aware Aggregation for Federated Learning
论文作者
论文摘要
联合学习(FL)使多个客户能够在不共享本地数据的情况下进行协作训练模型,并成为重要的隐私机器学习框架。但是,古典FL面临严重的安全性和鲁棒性问题,例如,恶意客户可以毒害模型更新,同时声称大量数量以扩大其模型更新在模型聚合中的影响。现有的FL防御方法,同时所有处理恶意模型更新的情况,请处理所有良性数量,或者简单地忽略/截断所有客户的数量。前者容易受到数量增强攻击的影响,而后者则导致次优性能,因为不同客户的本地数据通常具有明显不同的大小。在本文中,我们提出了一种称为Fedra的Federated学习的强大数量汇总算法,以意识到当地数据数量的意识,同时能够抵抗数量增强的攻击,以执行聚合。更具体地说,我们提出了一种通过共同考虑来自不同客户端的上传模型更新和数据数量的方法来过滤恶意客户端的方法,并在剩余客户端的模型更新中执行数量了解的加权平均。此外,随着参加联邦学习的恶意客户的数量可能会在不同的回合中动态变化,我们还提出了一个恶意的客户次数估算器,以预测每回合应过滤多少可疑客户。四个公共数据集上的实验证明了我们的联邦用方法在防御数量增强攻击方面的有效性。
Federated learning (FL) enables multiple clients to collaboratively train models without sharing their local data, and becomes an important privacy-preserving machine learning framework. However, classical FL faces serious security and robustness problem, e.g., malicious clients can poison model updates and at the same time claim large quantities to amplify the impact of their model updates in the model aggregation. Existing defense methods for FL, while all handling malicious model updates, either treat all quantities benign or simply ignore/truncate the quantities of all clients. The former is vulnerable to quantity-enhanced attack, while the latter leads to sub-optimal performance since the local data on different clients is usually in significantly different sizes. In this paper, we propose a robust quantity-aware aggregation algorithm for federated learning, called FedRA, to perform the aggregation with awareness of local data quantities while being able to defend against quantity-enhanced attacks. More specifically, we propose a method to filter malicious clients by jointly considering the uploaded model updates and data quantities from different clients, and performing quantity-aware weighted averaging on model updates from remaining clients. Moreover, as the number of malicious clients participating in the federated learning may dynamically change in different rounds, we also propose a malicious client number estimator to predict how many suspicious clients should be filtered in each round. Experiments on four public datasets demonstrate the effectiveness of our FedRA method in defending FL against quantity-enhanced attacks.