论文标题
沟通高效的随机零级优化用于联合学习
Communication-Efficient Stochastic Zeroth-Order Optimization for Federated Learning
论文作者
论文摘要
作为新兴的边缘人工智能范式,联合学习(FL)使许多边缘设备能够在不共享私人数据的情况下协作训练全球模型。为了提高FL的训练效率,已经提出了各种算法,从一阶到二阶方法。但是,这些算法不能应用于没有梯度信息的情况下,例如联合的黑盒攻击和联合的超参数调整。为了解决这个问题,在本文中,我们提出了一个无衍生的联合零订单优化(FEDZO)算法,该算法通过在每个通信回合中基于随机梯度估计器执行多个本地更新,并启用部分设备的参与。在非convex设置下,我们得出了FedZo算法在非独立且相同分布的数据上的收敛性能,并表征了本地迭代的数量和参与边缘设备对收敛的影响。为了使沟通高效的Fedzo超过无线网络,我们进一步提出了一项空中计算(AIRCOMP)辅助FEDZO算法。通过适当的收发器设计,我们表明,在某些信噪比条件下,仍然可以保留AirComp辅助FedZO的收敛性。仿真结果证明了FEDZO算法的有效性并验证了理论观察。
Federated learning (FL), as an emerging edge artificial intelligence paradigm, enables many edge devices to collaboratively train a global model without sharing their private data. To enhance the training efficiency of FL, various algorithms have been proposed, ranging from first-order to second-order methods. However, these algorithms cannot be applied in scenarios where the gradient information is not available, e.g., federated black-box attack and federated hyperparameter tuning. To address this issue, in this paper we propose a derivative-free federated zeroth-order optimization (FedZO) algorithm featured by performing multiple local updates based on stochastic gradient estimators in each communication round and enabling partial device participation. Under non-convex settings, we derive the convergence performance of the FedZO algorithm on non-independent and identically distributed data and characterize the impact of the numbers of local iterates and participating edge devices on the convergence. To enable communication-efficient FedZO over wireless networks, we further propose an over-the-air computation (AirComp) assisted FedZO algorithm. With an appropriate transceiver design, we show that the convergence of AirComp-assisted FedZO can still be preserved under certain signal-to-noise ratio conditions. Simulation results demonstrate the effectiveness of the FedZO algorithm and validate the theoretical observations.