论文标题
野生动物园:稀疏性使联盟学习有限和不可靠的沟通
SAFARI: Sparsity enabled Federated Learning with Limited and Unreliable Communications
论文作者
论文摘要
联合学习(FL)使边缘设备能够以分布式方式协作学习模型。许多现有研究的重点是提高高维模型的沟通效率,并解决由本地更新引起的偏见。但是,大多数FL算法要么基于可靠的通信,要么具有固定的和已知的不可靠特征。在实践中,网络可能会遭受动态的通道条件和非确定性破坏,并具有时间变化和未知的特征。为此,在本文中,我们提出了一种稀疏性,可以通过沟通效率和降低偏见为Safari,这使FL框架均被称为Safari。它使客户模型之间的相似性具有新颖的方式来纠正和弥补由不可靠的通信产生的偏见。更确切地说,在本地客户端实施了稀疏学习,以减轻沟通开销,而为了应对不可靠的通信,提出了一种基于相似性的补偿方法,以为缺失的模型更新提供替代物。我们在有限的差异和稀疏模型下分析了野生动物园。已经证明,在不可靠的通信下,Safari可以保证与具有完美通信的标准FedAvg相同的速度。对CIFAR-10数据集的实施和评估通过表明它可以实现与使用完美通信的FedAvg相同的收敛速度和准确性来验证Safari的有效性,其中多达80%的模型权重被修剪,并且每回合中缺少的客户更新很高。
Federated learning (FL) enables edge devices to collaboratively learn a model in a distributed fashion. Many existing researches have focused on improving communication efficiency of high-dimensional models and addressing bias caused by local updates. However, most of FL algorithms are either based on reliable communications or assume fixed and known unreliability characteristics. In practice, networks could suffer from dynamic channel conditions and non-deterministic disruptions, with time-varying and unknown characteristics. To this end, in this paper we propose a sparsity enabled FL framework with both communication efficiency and bias reduction, termed as SAFARI. It makes novel use of a similarity among client models to rectify and compensate for bias that is resulted from unreliable communications. More precisely, sparse learning is implemented on local clients to mitigate communication overhead, while to cope with unreliable communications, a similarity-based compensation method is proposed to provide surrogates for missing model updates. We analyze SAFARI under bounded dissimilarity and with respect to sparse models. It is demonstrated that SAFARI under unreliable communications is guaranteed to converge at the same rate as the standard FedAvg with perfect communications. Implementations and evaluations on CIFAR-10 dataset validate the effectiveness of SAFARI by showing that it can achieve the same convergence speed and accuracy as FedAvg with perfect communications, with up to 80% of the model weights being pruned and a high percentage of client updates missing in each round.