论文标题

FedRecover:使用历史信息从联邦学习中的中毒攻击中恢复

FedRecover: Recovering from Poisoning Attacks in Federated Learning using Historical Information

论文作者

Cao, Xiaoyu, Jia, Jinyuan, Zhang, Zaixi, Gong, Neil Zhenqiang

论文摘要

联邦学习很容易受到中毒攻击的影响,在这种攻击中,恶意客户通过向服务器发送恶意模型更新而毒害全球模型。现有的防御措施专注于防止少数恶意客户通过强大的联合学习方法中毒全球模型,并在有大量的情况下检测恶意客户。但是,在发现恶意客户之后,如何从中毒攻击中恢复全球模型仍然是一个公开挑战。一个天真的解决方案是删除检测到的恶意客户端并从头开始训练新的全球模型,这会导致巨大的成本,这对于智能手机和物联网设备等资源受限客户端可能无法忍受。 在这项工作中,我们提出了FedRecover,该公司可以从对客户的成本较小的中毒攻击中恢复准确的全球模型。我们的关键想法是,服务器估计客户端的模型更新,而不是要求客户在恢复过程中计算和通信。特别是,当训练中毒的全球模型时,服务器将在每回合中存储全局模型和客户端模型更新。在恢复过程中,服务器使用其存储的历史信息在每个回合中估算客户端的模型更新。此外,我们进一步优化了FedRecover,以使用热身,周期性更正,异常修复和最终调整策略恢复更准确的全局模型,在该策略中,服务器要求客户计算和传达其确切的模型更新。从理论上讲,我们表明,FedRecover恢复的全局模型与在某些假设下恢复的火车 - 划痕相同或相同。从经验上讲,我们对四个数据集的评估,三种联合学习方法,以及未靶向和有针对性的中毒攻击(例如,后门攻击)表明,FedRecover既准确又有效。

Federated learning is vulnerable to poisoning attacks in which malicious clients poison the global model via sending malicious model updates to the server. Existing defenses focus on preventing a small number of malicious clients from poisoning the global model via robust federated learning methods and detecting malicious clients when there are a large number of them. However, it is still an open challenge how to recover the global model from poisoning attacks after the malicious clients are detected. A naive solution is to remove the detected malicious clients and train a new global model from scratch, which incurs large cost that may be intolerable for resource-constrained clients such as smartphones and IoT devices. In this work, we propose FedRecover, which can recover an accurate global model from poisoning attacks with small cost for the clients. Our key idea is that the server estimates the clients' model updates instead of asking the clients to compute and communicate them during the recovery process. In particular, the server stores the global models and clients' model updates in each round, when training the poisoned global model. During the recovery process, the server estimates a client's model update in each round using its stored historical information. Moreover, we further optimize FedRecover to recover a more accurate global model using warm-up, periodic correction, abnormality fixing, and final tuning strategies, in which the server asks the clients to compute and communicate their exact model updates. Theoretically, we show that the global model recovered by FedRecover is close to or the same as that recovered by train-from-scratch under some assumptions. Empirically, our evaluation on four datasets, three federated learning methods, as well as untargeted and targeted poisoning attacks (e.g., backdoor attacks) shows that FedRecover is both accurate and efficient.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源