Abstract:
As a privacy-preserving distributed learning paradigm, federated learning (FL) has been proven to be vulnerable to various attacks, among which backdoor attack is one of ...Show MoreMetadata
Abstract:
As a privacy-preserving distributed learning paradigm, federated learning (FL) has been proven to be vulnerable to various attacks, among which backdoor attack is one of the toughest. In this attack, malicious users attempt to embed backdoor triggers into local models, resulting in the crafted inputs being misclassified as the targeted labels. To address such attack, several defense mechanisms are proposed, but may lose the effectiveness due to the following drawbacks. First, current methods heavily rely on massive labeled clean data, which is an impractical setting in FL. Moreover, an in-avoidable performance degradation usually occurs in the defensive procedure. To alleviate such concerns, we propose BadCleaner, a lossless and efficient backdoor defense scheme via attention-based federated multi-teacher distillation. First, BadCleaner can effectively tune the backdoored joint model without performance degradation, by distilling the in-depth knowledge from multiple teachers with only a small part of unlabeled clean data. Second, to fully eliminate the hidden backdoor patterns, we present an attention transfer method to alleviate the attention of models to the trigger regions. The extensive evaluation demonstrates that BadCleaner can reduce the success rates of state-of-the-art backdoor attacks without compromising the model performance.
Published in: IEEE Transactions on Dependable and Secure Computing ( Volume: 21, Issue: 5, Sept.-Oct. 2024)
Funding Agency:
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Federated Learning ,
- Backdoor Attacks ,
- Attacks In Federated Learning ,
- Model Performance ,
- Clean Data ,
- Unlabeled Data ,
- Joint Model ,
- Target Label ,
- Multiple Teachers ,
- Malicious Users ,
- Normal Samples ,
- Global Model ,
- Intermediate Layer ,
- Channel Dimension ,
- Student Model ,
- Attention Map ,
- Central Server ,
- Threat Model ,
- Student Network ,
- Local Updates ,
- Attack Success Rate ,
- Clean Dataset ,
- Defense Methods ,
- Pixel Block ,
- High Model Performance ,
- Voting Scheme ,
- Attention Loss ,
- Federated Learning Framework ,
- Aggregation Algorithm ,
- Communication Rounds
- Author Keywords
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Federated Learning ,
- Backdoor Attacks ,
- Attacks In Federated Learning ,
- Model Performance ,
- Clean Data ,
- Unlabeled Data ,
- Joint Model ,
- Target Label ,
- Multiple Teachers ,
- Malicious Users ,
- Normal Samples ,
- Global Model ,
- Intermediate Layer ,
- Channel Dimension ,
- Student Model ,
- Attention Map ,
- Central Server ,
- Threat Model ,
- Student Network ,
- Local Updates ,
- Attack Success Rate ,
- Clean Dataset ,
- Defense Methods ,
- Pixel Block ,
- High Model Performance ,
- Voting Scheme ,
- Attention Loss ,
- Federated Learning Framework ,
- Aggregation Algorithm ,
- Communication Rounds
- Author Keywords