Mitigating Backdoor Attacks in Federated Learning Using PPA and MiniMax Game Theory
AI 摘要
该论文提出FedBBA,利用信誉系统、激励机制和博弈论方法,减轻联邦学习中后门攻击的影响。
主要贡献
- 提出FedBBA框架,结合信誉系统、激励机制和博弈论
- 使用PPA和MiniMax博弈论动态识别并最小化恶意客户端的影响
- 在GTSRB和BTSC数据集上验证了FedBBA的有效性
方法论
结合信誉系统评估客户端行为,激励诚实参与,惩罚恶意行为,并使用PPA和博弈论减少恶意影响。
原文摘要
Federated Learning (FL) is witnessing wider adoption due to its ability to benefit from large amounts of scattered data while preserving privacy. However, despite its advantages, federated learning suffers from several setbacks that directly impact the accuracy, and the integrity of the global model it produces. One of these setbacks is the presence of malicious clients who actively try to harm the global model by injecting backdoor data into their local models while trying to evade detection. The objective of such clients is to trick the global model into making false predictions during inference, thereby compromising the integrity and trustworthiness of the global model on which honest stakeholders rely. To mitigate such mischievous behavior, we propose FedBBA (Federated Backdoor and Behavior Analysis). The proposed model aims to dampen the effect of such clients on the final accuracy, creating more resilient federated learning environments. We engineer our approach through the combination of (1) a reputation system to evaluate and track client behavior, (2) an incentive mechanism to reward honest participation and penalize malicious behavior, and (3) game theoretical models with projection pursuit analysis (PPA) to dynamically identify and minimize the impact of malicious clients on the global model. Extensive simulations on the German Traffic Sign Recognition Benchmark (GTSRB) and Belgium Traffic Sign Classification (BTSC) datasets demonstrate that FedBBA reduces the backdoor attack success rate to approximately 1.1%--11% across various attack scenarios, significantly outperforming state-of-the-art defenses like RDFL and RoPE, which yielded attack success rates between 23% and 76%, while maintaining high normal task accuracy (~95%--98%).