重新思考规模:智能体范式下小语言模型的部署权衡 / Rethinking Scale: Deployment Trade-offs of Small Language Models under Agent Paradigms
1️⃣ 一句话总结
本文系统比较了小于100亿参数的小语言模型在三种不同模式(基础模型、单智能体使用工具、多智能体协作)下的性能与成本,发现单智能体系统在部署时能达到最佳的效率与效果平衡,而多智能体模式虽能提升能力但会带来额外的计算开销。
Despite the impressive capabilities of large language models, their substantial computational costs, latency, and privacy risks hinder their widespread deployment in real-world applications. Small Language Models (SLMs) with fewer than 10 billion parameters present a promising alternative; however, their inherent limitations in knowledge and reasoning curtail their effectiveness. Existing research primarily focuses on enhancing SLMs through scaling laws or fine-tuning strategies while overlooking the potential of using agent paradigms, such as tool use and multi-agent collaboration, to systematically compensate for the inherent weaknesses of small models. To address this gap, this paper presents the first large-scale, comprehensive study of <10B open-source models under three paradigms: (1) the base model, (2) a single agent equipped with tools, and (3) a multi-agent system with collaborative capabilities. Our results show that single-agent systems achieve the best balance between performance and cost, while multi-agent setups add overhead with limited gains. Our findings highlight the importance of agent-centric design for efficient and trustworthy deployment in resource-constrained settings.
重新思考规模:智能体范式下小语言模型的部署权衡 / Rethinking Scale: Deployment Trade-offs of Small Language Models under Agent Paradigms
本文系统比较了小于100亿参数的小语言模型在三种不同模式(基础模型、单智能体使用工具、多智能体协作)下的性能与成本,发现单智能体系统在部署时能达到最佳的效率与效果平衡,而多智能体模式虽能提升能力但会带来额外的计算开销。
源自 arXiv: 2604.19299