菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-09
📄 Abstract - SDFed: Bridging Local Global Discrepancy via Subspace Refinement and Divergence Control in Federated Prompt Learning

Vision-language pretrained models offer strong transferable representations, yet adapting them in privacy-sensitive multi-party settings is challenging due to the high communication cost of federated optimization and the limited local data on clients. Federated prompt learning mitigates this issue by keeping the VLPM backbone frozen and collaboratively training lightweight prompt parameters. However, existing approaches typically enforce a unified prompt structure and length across clients, which is inadequate under practical client heterogeneity in both data distributions and system resources, and may further introduce conflicts between globally shared and locally optimal knowledge. To address these challenges, we propose \textbf{SDFed}, a heterogeneous federated prompt learning framework that bridges Local-Global Discrepancy via Subspace Refinement and Divergence Control. SDFed maintains a fixed-length global prompt for efficient aggregation while allowing each client to learn a variable-length local prompt to better match its data characteristics and capacity. To mitigate local-global conflicts and facilitate effective knowledge transfer, SDFed introduces a subspace refinement method for local prompts and an information retention and divergence control strategy that preserves key local information while maintaining appropriate separability between global and local representations. Extensive experiments on several datasets demonstrate that SDFed consistently improves performance and robustness in heterogeneous federated settings.

顶级标签: multi-modal model training machine learning
详细标签: federated learning prompt learning vision-language models heterogeneous clients privacy 或 搜索:

SDFed:通过子空间精炼与差异控制解决联邦提示学习中的本地-全局差异问题 / SDFed: Bridging Local Global Discrepancy via Subspace Refinement and Divergence Control in Federated Prompt Learning


1️⃣ 一句话总结

这篇论文提出了一种名为SDFed的新方法,它允许不同设备在保护隐私的联合训练中,根据自身数据特点和计算能力灵活调整学习参数,并通过精巧的设计减少局部优化与全局共享知识之间的冲突,从而在数据分布和设备资源不均的现实场景下,显著提升了视觉-语言大模型的适应效果和鲁棒性。

源自 arXiv: 2602.08590