菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-16
📄 Abstract - MA-VLCM: A Vision Language Critic Model for Value Estimation of Policies in Multi-Agent Team Settings

Multi-agent reinforcement learning (MARL) commonly relies on a centralized critic to estimate the value function. However, learning such a critic from scratch is highly sample-inefficient and often lacks generalization across environments. At the same time, large vision-language-action models (VLAs) trained on internet-scale data exhibit strong multimodal reasoning and zero-shot generalization capabilities, yet directly deploying them for robotic execution remains computationally prohibitive, particularly in heterogeneous multi-robot systems with diverse embodiments and resource constraints. To address these challenges, we propose Multi-Agent Vision-Language-Critic Models (MA-VLCM), a framework that replaces the learned centralized critic in MARL with a pretrained vision-language model fine-tuned to evaluate multi-agent behavior. MA-VLCM acts as a centralized critic conditioned on natural language task descriptions, visual trajectory observations, and structured multi-agent state information. By eliminating critic learning during policy optimization, our approach significantly improves sample efficiency while producing compact execution policies suitable for deployment on resource-constrained robots. Results show good zero-shot return estimation on models with differing VLM backbones on in-distribution and out-of-distribution scenarios in multi-agent team settings

顶级标签: multi-modal agents reinforcement learning
详细标签: multi-agent reinforcement learning vision-language model value estimation sample efficiency centralized critic 或 搜索:

MA-VLCM:一种用于多智能体团队场景中策略价值评估的视觉语言批评模型 / MA-VLCM: A Vision Language Critic Model for Value Estimation of Policies in Multi-Agent Team Settings


1️⃣ 一句话总结

这篇论文提出了一种新方法,利用预训练好的视觉语言大模型来快速评估多机器人团队的合作表现,从而大幅减少了训练所需的数据量,并能让训练好的策略直接部署在计算资源有限的真实机器人上。

源自 arXiv: 2603.15418