论文标题
反转梯度 - 在联邦学习中打破隐私有多容易?
Inverting Gradients -- How easy is it to break privacy in federated learning?
论文作者
论文摘要
联合学习的想法是在服务器上协作培训神经网络。每个用户都会收到网络的当前权重,并且根据本地数据转动发送参数更新(梯度)。该协议的设计不仅是为了训练神经网络有效地培训数据,而且还为用户提供隐私福利,因为其输入数据保留在设备上,并且仅共享参数梯度。但是共享参数梯度的安全性如何?以前的攻击仅通过在人为的设置中取得成功,即使是单个图像,也提供了一种错误的安全感。但是,通过利用幅度不变的损失以及基于对抗性攻击的优化策略,我们表明实际上可以从高分辨率的参数梯度中忠实地重建图像,并证明即使是训练有素的深网,也可以证明这种隐私也可能破坏。我们分析了体系结构以及参数对重建输入图像的难度的影响,并证明对完全连接的层的任何输入都可以在分析上与其余体系结构无关。最后,我们讨论了在实践中遇到的设置,并表明即使在几个迭代或几个图像上平均梯度也不能保护用户在计算机视觉中联合学习应用程序中的隐私。
The idea of federated learning is to collaboratively train a neural network on a server. Each user receives the current weights of the network and in turns sends parameter updates (gradients) based on local data. This protocol has been designed not only to train neural networks data-efficiently, but also to provide privacy benefits for users, as their input data remains on device and only parameter gradients are shared. But how secure is sharing parameter gradients? Previous attacks have provided a false sense of security, by succeeding only in contrived settings - even for a single image. However, by exploiting a magnitude-invariant loss along with optimization strategies based on adversarial attacks, we show that is is actually possible to faithfully reconstruct images at high resolution from the knowledge of their parameter gradients, and demonstrate that such a break of privacy is possible even for trained deep networks. We analyze the effects of architecture as well as parameters on the difficulty of reconstructing an input image and prove that any input to a fully connected layer can be reconstructed analytically independent of the remaining architecture. Finally we discuss settings encountered in practice and show that even averaging gradients over several iterations or several images does not protect the user's privacy in federated learning applications in computer vision.