菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-31
📄 Abstract - Beyond Corner Patches: Semantics-Aware Backdoor Attack in Federated Learning

Backdoor attacks on federated learning (FL) are most often evaluated with synthetic corner patches or out-of-distribution (OOD) patterns that are unlikely to arise in practice. In this paper, we revisit the backdoor threat to standard FL (a single global model) under a more realistic setting where triggers must be semantically meaningful, in-distribution, and visually plausible. We propose SABLE, a Semantics-Aware Backdoor for LEarning in federated settings, which constructs natural, content-consistent triggers (e.g., semantic attribute changes such as sunglasses) and optimizes an aggregation-aware malicious objective with feature separation and parameter regularization to keep attacker updates close to benign ones. We instantiate SABLE on CelebA hair-color classification and the German Traffic Sign Recognition Benchmark (GTSRB), poisoning only a small, interpretable subset of each malicious client's local data while otherwise following the standard FL protocol. Across heterogeneous client partitions and multiple aggregation rules (FedAvg, Trimmed Mean, MultiKrum, and FLAME), our semantics-driven triggers achieve high targeted attack success rates while preserving benign test accuracy. These results show that semantics-aligned backdoors remain a potent and practical threat in federated learning, and that robustness claims based solely on synthetic patch triggers can be overly optimistic.

顶级标签: systems model training machine learning
详细标签: federated learning backdoor attack adversarial robustness semantic triggers security 或 搜索:

超越角落补丁:联邦学习中的语义感知后门攻击 / Beyond Corner Patches: Semantics-Aware Backdoor Attack in Federated Learning


1️⃣ 一句话总结

这篇论文提出了一种名为SABLE的新型后门攻击方法,它利用语义上自然且合理的触发器(如给人像添加太阳镜),在联邦学习中隐蔽地植入后门,即使在多种防御聚合规则下也能保持高攻击成功率,揭示了基于合成补丁的防御评估可能过于乐观。

源自 arXiv: 2603.29328