{"title":"Toward Efficient and Certified Recovery From Poisoning Attacks in Federated Learning","authors":"Yu Jiang;Jiyuan Shen;Ziyao Liu;Chee Wei Tan;Kwok-Yan Lam","doi":"10.1109/TIFS.2025.3533907","DOIUrl":null,"url":null,"abstract":"Federated learning (FL) is vulnerable to poisoning attacks, where malicious clients manipulate their updates to affect the global model. Although various methods exist for detecting such clients in FL, identifying malicious clients requires sufficient model updates, and hence by the time malicious clients are detected, FL models have already been poisoned. Thus, a method is needed to recover an accurate global model after malicious clients are identified. Current recovery methods rely on (i) all historical information from participating FL clients and (ii) the initial model unaffected by the malicious clients, both leading to a high demand for storage and computational resources. In this paper, we show that highly effective recovery can still be achieved based on 1) selective historical information rather than all historical information and 2) a historical model that has not been significantly affected by malicious clients rather than the initial model. In this scenario, we can accelerate the recovery speed and decrease memory consumption while maintaining comparable recovery performance. Following this concept, we introduce Crab (Certified Recovery from Poisoning Attacks and Breaches), an efficient and certified recovery method, which relies on selective information storage and adaptive model rollback. Theoretically, we demonstrate that the difference between the global model recovered by Crab and the one recovered by train-from-scratch can be bounded under certain assumptions. Our experiments, performed across four datasets with multiple machine learning models and aggregation methods, involving both untargeted and targeted poisoning attacks, demonstrate that Crab is not only accurate and efficient but also consistently outperforms previous approaches in recovery speed and memory consumption.","PeriodicalId":13492,"journal":{"name":"IEEE Transactions on Information Forensics and Security","volume":"20 ","pages":"2632-2647"},"PeriodicalIF":8.0000,"publicationDate":"2025-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Information Forensics and Security","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10852413/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0
Abstract
Federated learning (FL) is vulnerable to poisoning attacks, where malicious clients manipulate their updates to affect the global model. Although various methods exist for detecting such clients in FL, identifying malicious clients requires sufficient model updates, and hence by the time malicious clients are detected, FL models have already been poisoned. Thus, a method is needed to recover an accurate global model after malicious clients are identified. Current recovery methods rely on (i) all historical information from participating FL clients and (ii) the initial model unaffected by the malicious clients, both leading to a high demand for storage and computational resources. In this paper, we show that highly effective recovery can still be achieved based on 1) selective historical information rather than all historical information and 2) a historical model that has not been significantly affected by malicious clients rather than the initial model. In this scenario, we can accelerate the recovery speed and decrease memory consumption while maintaining comparable recovery performance. Following this concept, we introduce Crab (Certified Recovery from Poisoning Attacks and Breaches), an efficient and certified recovery method, which relies on selective information storage and adaptive model rollback. Theoretically, we demonstrate that the difference between the global model recovered by Crab and the one recovered by train-from-scratch can be bounded under certain assumptions. Our experiments, performed across four datasets with multiple machine learning models and aggregation methods, involving both untargeted and targeted poisoning attacks, demonstrate that Crab is not only accurate and efficient but also consistently outperforms previous approaches in recovery speed and memory consumption.
联邦学习(FL)很容易受到中毒攻击,在这种攻击中,恶意客户端操纵它们的更新来影响全局模型。虽然在FL中存在各种检测此类客户端的方法,但识别恶意客户端需要足够的模型更新,因此当检测到恶意客户端时,FL模型已经中毒。因此,需要一种在识别出恶意客户端后恢复准确全局模型的方法。当前的恢复方法依赖于(i)来自参与的FL客户机的所有历史信息和(ii)不受恶意客户机影响的初始模型,这两者都导致对存储和计算资源的高需求。在本文中,我们证明了高效的恢复仍然可以基于1)选择性历史信息而不是所有历史信息和2)没有受到恶意客户端显著影响的历史模型而不是初始模型来实现。在这种情况下,我们可以加快恢复速度并减少内存消耗,同时保持相当的恢复性能。根据这一概念,我们介绍了Crab (Certified Recovery from Poisoning Attacks and breach),这是一种高效的认证恢复方法,它依赖于选择性信息存储和自适应模型回滚。从理论上讲,我们证明了在一定的假设下,螃蟹恢复的全局模型与从头开始的火车恢复的全局模型之间的差异是有界的。我们在四个数据集上使用多种机器学习模型和聚合方法进行了实验,包括非靶向和靶向中毒攻击,结果表明Crab不仅准确高效,而且在恢复速度和内存消耗方面始终优于以前的方法。
期刊介绍:
The IEEE Transactions on Information Forensics and Security covers the sciences, technologies, and applications relating to information forensics, information security, biometrics, surveillance and systems applications that incorporate these features