论文标题

逐步逐步减轻的位置异质联邦学习

Bitwidth Heterogeneous Federated Learning with Progressive Weight Dequantization

论文作者

Yoon, Jaehong, Park, Geon, Jeong, Wonyong, Hwang, Sung Ju

论文摘要

在实用的联合学习方案中,参与的设备可能具有不同的位宽,用于按设计进行计算和内存存储。然而,尽管设备异质联合学习方案取得了进展,但硬件中的位 - 位规范的异质性大多被忽略了。我们介绍了一种务实的FL场景,在参与设备中具有位于位的异质性,被称为Bitwidth异质联邦学习(BHFL)。 BHFL带来了一个新的挑战,即具有不同位宽度的模型参数的聚合可能会导致严重的性能变性,尤其是对于高宽度模型。为了解决这个问题,我们提出了ProWD框架,该框架在中央服务器上具有可训练的权重去除剂,该框架逐渐将低位宽度重建为更高的位宽度重量,最后将其重建为完全精确的重量。 PROWD进一步选择性地汇总了模型参数,以最大程度地提高位异质权重的兼容性。我们使用具有变化的位宽的客户端在基准数据集上的相关FL基准验证了Prowd。我们的prowd在很大程度上胜过基线FL算法以及在拟议的BHFL方案下的天真方法(例如,平均分组)。

In practical federated learning scenarios, the participating devices may have different bitwidths for computation and memory storage by design. However, despite the progress made in device-heterogeneous federated learning scenarios, the heterogeneity in the bitwidth specifications in the hardware has been mostly overlooked. We introduce a pragmatic FL scenario with bitwidth heterogeneity across the participating devices, dubbed as Bitwidth Heterogeneous Federated Learning (BHFL). BHFL brings in a new challenge, that the aggregation of model parameters with different bitwidths could result in severe performance degeneration, especially for high-bitwidth models. To tackle this problem, we propose ProWD framework, which has a trainable weight dequantizer at the central server that progressively reconstructs the low-bitwidth weights into higher bitwidth weights, and finally into full-precision weights. ProWD further selectively aggregates the model parameters to maximize the compatibility across bit-heterogeneous weights. We validate ProWD against relevant FL baselines on the benchmark datasets, using clients with varying bitwidths. Our ProWD largely outperforms the baseline FL algorithms as well as naive approaches (e.g. grouped averaging) under the proposed BHFL scenario.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源