论文标题

与部分客户参与的联合学习的锚定抽样

Anchor Sampling for Federated Learning with Partial Client Participation

论文作者

Wu, Feijie, Guo, Song, Qu, Zhihao, He, Shiqi, Liu, Ziming, Gao, Jing

论文摘要

与全面的客户参与相比,部分客户参与是联合学习中更实用的情况,但它可能会扩大联盟学习中的一些挑战,例如数据异质性。缺乏部分客户参与的不活动客户的更新使模型聚合更有可能基于完整的客户参与而偏离聚合。提议对各个客户进行大量培训,以解决一般数据异质性,但他们在部分客户参与下的有效性尚不清楚。在这些挑战中,我们建议开发一个新颖的联邦学习框架,称为FedAmd,以部分客户参与。核心思想是锚定抽样,将部分参与者分为锚和矿工群体。锚点组中的每个客户均使用大批量计算梯度计算的瞄准本地靶心。在Bullseyes的指导下,矿工组中的客户使用小批次转向多个近乎最佳的本地更新,并更新全球模型。通过整合两组的结果,FedAmd能够加速训练过程并改善模型性能。通过$ε$ -Approximation衡量并与最先进的方法进行了比较,FedAmd在非凸面目标下最多可达到$ O(1/ε)$(1/ε)$更少的通信回合。关于现实世界数据集的实证研究验证了FedAmd的有效性,并证明了所提出的算法的优越性:它不仅可以大大节省计算和通信成本,而且还可以显着提高测试准确性。

Compared with full client participation, partial client participation is a more practical scenario in federated learning, but it may amplify some challenges in federated learning, such as data heterogeneity. The lack of inactive clients' updates in partial client participation makes it more likely for the model aggregation to deviate from the aggregation based on full client participation. Training with large batches on individual clients is proposed to address data heterogeneity in general, but their effectiveness under partial client participation is not clear. Motivated by these challenges, we propose to develop a novel federated learning framework, referred to as FedAMD, for partial client participation. The core idea is anchor sampling, which separates partial participants into anchor and miner groups. Each client in the anchor group aims at the local bullseye with the gradient computation using a large batch. Guided by the bullseyes, clients in the miner group steer multiple near-optimal local updates using small batches and update the global model. By integrating the results of the two groups, FedAMD is able to accelerate the training process and improve the model performance. Measured by $ε$-approximation and compared to the state-of-the-art methods, FedAMD achieves the convergence by up to $O(1/ε)$ fewer communication rounds under non-convex objectives. Empirical studies on real-world datasets validate the effectiveness of FedAMD and demonstrate the superiority of the proposed algorithm: Not only does it considerably save computation and communication costs, but also the test accuracy significantly improves.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源