菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-29
📄 Abstract - SplitFT: An Adaptive Federated Split Learning System For LLMs Fine-Tuning

Federated Split Learning has been identified as an efficient approach to address the computational resource constraints of clients in classical federated learning, while guaranteeing data privacy for distributed model training across data owners. However, it faces some critical challenges when such a training strategy meets large language models (LLMs) for fine-tuning. Such challenges include setting the cutlayer adaptively across different clients to address the data and device heterogeneity issues, which affect the system performance significantly. In addition, efficiently reducing the communication overhead during the fine-tuning procedure is also another challenge. No work tries to address these challenges. To bridge this gap, we propose SplitTF, an adaptive federated split learning system for LLMs fine-tuning. SplitFT enables different clients to set different cut layers according to their computation resources and trained model performance. SplitFT also proposes to reduce the LoRA rank in cutlayer to reduce the communication overhead. In addition to simulating the heterogeneous data in real-world applications for our proposed split federated learning system, we propose a length-based Dirichlet approach to divide the training data into different clients. Extensive experimental results show that our proposed approach outperforms the state-of-the-art approach for fine-tuning time efficiency and model performance based on various popular benchmarks.

顶级标签: llm systems
详细标签: federated learning split learning fine-tuning communication efficiency heterogeneity 或 搜索:

SplitFT:一种自适应的联邦拆分学习系统用于大语言模型微调 / SplitFT: An Adaptive Federated Split Learning System For LLMs Fine-Tuning


1️⃣ 一句话总结

本文提出了一种名为SplitFT的自适应联邦拆分学习系统,能够让不同计算能力的客户端灵活选择模型拆分层,并通过减少拆分层的LoRA秩来降低通信开销,从而高效地微调大语言模型,同时保护数据隐私。

源自 arXiv: 2604.26388