菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-21
📄 Abstract - FedProxy: Federated Fine-Tuning of LLMs via Proxy SLMs and Heterogeneity-Aware Fusion

Federated fine-tuning of Large Language Models (LLMs) is obstructed by a trilemma of challenges: protecting LLMs intellectual property (IP), ensuring client privacy, and mitigating performance loss on heterogeneous data. Existing methods like Offsite-Tuning (OT) secure the LLMs IP by having clients train only lightweight adapters, yet our analysis reveals they suffer from a fundamental performance bottleneck, leaving a significant gap compared to centralized training. To bridge this gap, we introduce FedProxy, a new federated adaptation framework. FedProxy replaces weak adapters with a unified, powerful Proxy Small Language Model (SLM), compressed from the proprietary LLM, to serve as a high-fidelity surrogate for collaborative fine-tuning. Our framework systematically resolves the trilemma through a three-stage architecture: (i) Efficient Representation via server-guided compression to create a resource-friendly proxy; (ii) Robust Optimization through an interference-mitigating aggregation strategy to handle data heterogeneity; and (iii) Effortless Fusion via a training-free "plug-in" mechanism to integrate learned knowledge back into the LLM. Experiments show FedProxy significantly outperforms OT methods and approaches centralized performance, establishing a new benchmark for secure and high-performance federated LLM adaptation.

顶级标签: llm machine learning systems
详细标签: federated learning fine-tuning proxy slm heterogeneity-aware fusion intellectual property 或 搜索:

FedProxy:通过代理小语言模型和异构感知融合实现大语言模型的联邦微调 / FedProxy: Federated Fine-Tuning of LLMs via Proxy SLMs and Heterogeneity-Aware Fusion


1️⃣ 一句话总结

本文提出FedProxy框架,用一个从大模型压缩而来的小语言模型作为代理,在保护大模型知识产权和客户端隐私的前提下,通过三阶段架构高效解决联邦微调中性能下降与数据异构的难题,性能接近集中式训练。

源自 arXiv: 2604.19015