论文标题
通过有损分布式来源编码的联合学习:分析和优化
Federated Learning with Lossy Distributed Source Coding: Analysis and Optimization
论文作者
论文摘要
最近,将数据共享用模型共享取代的联邦学习(FL)已成为一种有效且对隐私友好的机器学习(ML)范式。 FL中的主要挑战之一是模型聚合的巨大沟通成本。已经提出了许多压缩/量化方案,以降低模型聚合的通信成本。但是,以下问题仍未得到答案:通信成本与FL融合绩效之间的基本权衡是什么?在本文中,我们设法回答了这个问题。具体而言,我们首先根据速率理论提出了模型聚合性能分析的一般框架。在提出的分析框架下,我们得出了模型聚合的速率延伸区域的内部结合。然后,我们进行FL收敛分析以连接聚集失真和FL收敛性能。我们提出一个聚集失真最小化问题,以提高FL收敛性能。开发了两种算法来解决上述问题。聚合失真,收敛性能和通信成本的数值结果表明,基线模型聚合方案仍然具有进一步改进的巨大潜力。
Recently, federated learning (FL), which replaces data sharing with model sharing, has emerged as an efficient and privacy-friendly machine learning (ML) paradigm. One of the main challenges in FL is the huge communication cost for model aggregation. Many compression/quantization schemes have been proposed to reduce the communication cost for model aggregation. However, the following question remains unanswered: What is the fundamental trade-off between the communication cost and the FL convergence performance? In this paper, we manage to answer this question. Specifically, we first put forth a general framework for model aggregation performance analysis based on the rate-distortion theory. Under the proposed analysis framework, we derive an inner bound of the rate-distortion region of model aggregation. We then conduct an FL convergence analysis to connect the aggregation distortion and the FL convergence performance. We formulate an aggregation distortion minimization problem to improve the FL convergence performance. Two algorithms are developed to solve the above problem. Numerical results on aggregation distortion, convergence performance, and communication cost demonstrate that the baseline model aggregation schemes still have great potential for further improvement.