论文标题
人群:联合后门检测
CrowdGuard: Federated Backdoor Detection in Federated Learning
论文作者
论文摘要
联邦学习(FL)是一种有前途的方法,使多个客户能够协作培训深度神经网络(DNN),而无需共享本地培训数据。但是,FL容易受到后门(或目标中毒)攻击的影响。这些攻击是由恶意客户发起的,他们试图通过将特定的行为引入学习模型中,以妥协被精心制作的投入触发。现有的FL保障措施有各种局限性:它们仅限于特定的数据分布或由于排除良性模型或添加噪声而降低全局模型的精度,容易受到自适应防御感知对手的影响,或者要求服务器访问本地模型,从而允许数据推荐攻击。 本文提出了一种新颖的防御机制,即Crowdguard,该机制有效地减轻了FL中的后门攻击,并克服了现有技术的缺陷。它利用客户对单个模型的反馈,分析隐藏层中神经元的行为,并通过迭代修剪方案消除了中毒的模型。 CrowdGuard采用服务器上堆叠的聚类方案来增强其对Rogue客户反馈的韧性。评估结果表明,CrowdGuard在各种情况(包括IID和非IID数据分布)中实现了100%的真实利率和真实负率。此外,人群守卫承受自适应对手,同时保留了受保护模型的原始性能。为了确保机密性,CrowdGuard使用安全且保存隐私的架构,利用客户和服务器侧的可信赖执行环境(TEE)。
Federated Learning (FL) is a promising approach enabling multiple clients to train Deep Neural Networks (DNNs) collaboratively without sharing their local training data. However, FL is susceptible to backdoor (or targeted poisoning) attacks. These attacks are initiated by malicious clients who seek to compromise the learning process by introducing specific behaviors into the learned model that can be triggered by carefully crafted inputs. Existing FL safeguards have various limitations: They are restricted to specific data distributions or reduce the global model accuracy due to excluding benign models or adding noise, are vulnerable to adaptive defense-aware adversaries, or require the server to access local models, allowing data inference attacks. This paper presents a novel defense mechanism, CrowdGuard, that effectively mitigates backdoor attacks in FL and overcomes the deficiencies of existing techniques. It leverages clients' feedback on individual models, analyzes the behavior of neurons in hidden layers, and eliminates poisoned models through an iterative pruning scheme. CrowdGuard employs a server-located stacked clustering scheme to enhance its resilience to rogue client feedback. The evaluation results demonstrate that CrowdGuard achieves a 100% True-Positive-Rate and True-Negative-Rate across various scenarios, including IID and non-IID data distributions. Additionally, CrowdGuard withstands adaptive adversaries while preserving the original performance of protected models. To ensure confidentiality, CrowdGuard uses a secure and privacy-preserving architecture leveraging Trusted Execution Environments (TEEs) on both client and server sides.