FedProxy:通过代理小语言模型和异构感知融合实现大语言模型的联邦微调 / FedProxy: Federated Fine-Tuning of LLMs via Proxy SLMs and Heterogeneity-Aware Fusion
1️⃣ 一句话总结
本文提出FedProxy框架,用一个从大模型压缩而来的小语言模型作为代理,在保护大模型知识产权和客户端隐私的前提下,通过三阶段架构高效解决联邦微调中性能下降与数据异构的难题,性能接近集中式训练。
Federated fine-tuning of Large Language Models (LLMs) is obstructed by a trilemma of challenges: protecting LLMs intellectual property (IP), ensuring client privacy, and mitigating performance loss on heterogeneous data. Existing methods like Offsite-Tuning (OT) secure the LLMs IP by having clients train only lightweight adapters, yet our analysis reveals they suffer from a fundamental performance bottleneck, leaving a significant gap compared to centralized training. To bridge this gap, we introduce FedProxy, a new federated adaptation framework. FedProxy replaces weak adapters with a unified, powerful Proxy Small Language Model (SLM), compressed from the proprietary LLM, to serve as a high-fidelity surrogate for collaborative fine-tuning. Our framework systematically resolves the trilemma through a three-stage architecture: (i) Efficient Representation via server-guided compression to create a resource-friendly proxy; (ii) Robust Optimization through an interference-mitigating aggregation strategy to handle data heterogeneity; and (iii) Effortless Fusion via a training-free "plug-in" mechanism to integrate learned knowledge back into the LLM. Experiments show FedProxy significantly outperforms OT methods and approaches centralized performance, establishing a new benchmark for secure and high-performance federated LLM adaptation.
FedProxy:通过代理小语言模型和异构感知融合实现大语言模型的联邦微调 / FedProxy: Federated Fine-Tuning of LLMs via Proxy SLMs and Heterogeneity-Aware Fusion
本文提出FedProxy框架,用一个从大模型压缩而来的小语言模型作为代理,在保护大模型知识产权和客户端隐私的前提下,通过三阶段架构高效解决联邦微调中性能下降与数据异构的难题,性能接近集中式训练。
源自 arXiv: 2604.19015