论文标题

一个联合的深度学习框架,用于隐私保护和沟通效率

A Federated Deep Learning Framework for Privacy Preservation and Communication Efficiency

论文作者

Cao, Tien-Dung, Truong-Huu, Tram, Tran, Hien, Tran, Khanh

论文摘要

深度学习在许多应用中取得了巨大的成功。但是,其实践中的部署受到了两个问题的困扰:由于通常在地理位置上分布的大量数据,由于传输了大量数据,因此必须集中汇总的数据隐私和高通信开销。解决这两个问题都是具有挑战性的,大多数现有作品无法提供有效的解决方案。在本文中,我们开发了FedPC,这是一个联合的深度学习框架,以保护隐私和沟通效率。该框架允许在多个私人数据集上学习模型,同时也没有透露培训数据的任何信息,即使使用中间数据。该框架还可以最大程度地减少交换以更新模型的数据量。我们正式证明了与FEDPC培训及其隐私权的培训时学习模型的融合。我们进行了广泛的实验,以评估FEDPC的性能,以对上限性能(集中训练)和开销进行交流。结果表明,当数据分配到10个计算节点时,FEDPC将模型的性能近似保持在中央训练的$ 8.5 \%$之内。与现有作品相比,FEDPC还将通信开销降低了$ 42.20 \%$。

Deep learning has achieved great success in many applications. However, its deployment in practice has been hurdled by two issues: the privacy of data that has to be aggregated centrally for model training and high communication overhead due to transmission of a large amount of data usually geographically distributed. Addressing both issues is challenging and most existing works could not provide an efficient solution. In this paper, we develop FedPC, a Federated Deep Learning Framework for Privacy Preservation and Communication Efficiency. The framework allows a model to be learned on multiple private datasets while not revealing any information of training data, even with intermediate data. The framework also minimizes the amount of data exchanged to update the model. We formally prove the convergence of the learning model when training with FedPC and its privacy-preserving property. We perform extensive experiments to evaluate the performance of FedPC in terms of the approximation to the upper-bound performance (when training centrally) and communication overhead. The results show that FedPC maintains the performance approximation of the models within $8.5\%$ of the centrally-trained models when data is distributed to 10 computing nodes. FedPC also reduces the communication overhead by up to $42.20\%$ compared to existing works.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源