菜单

🤖 系统
📄 Abstract - Grounding Computer Use Agents on Human Demonstrations

Building reliable computer-use agents requires grounding: accurately connecting natural language instructions to the correct on-screen elements. While large datasets exist for web and mobile interactions, high-quality resources for desktop environments are limited. To address this gap, we introduce GroundCUA, a large-scale desktop grounding dataset built from expert human demonstrations. It covers 87 applications across 12 categories and includes 56K screenshots, with every on-screen element carefully annotated for a total of over 3.56M human-verified annotations. From these demonstrations, we generate diverse instructions that capture a wide range of real-world tasks, providing high-quality data for model training. Using GroundCUA, we develop the GroundNext family of models that map instructions to their target UI elements. At both 3B and 7B scales, GroundNext achieves state-of-the-art results across five benchmarks using supervised fine-tuning, while requiring less than one-tenth the training data of prior work. Reinforcement learning post-training further improves performance, and when evaluated in an agentic setting on the OSWorld benchmark using o3 as planner, GroundNext attains comparable or superior results to models trained with substantially more data,. These results demonstrate the critical role of high-quality, expert-driven datasets in advancing general-purpose computer-use agents.

顶级标签: agents systems model training
详细标签: computer-use agents ui grounding desktop automation human demonstrations instruction grounding 或 搜索:

📄 论文总结

基于人类演示的计算机使用智能体基础构建 / Grounding Computer Use Agents on Human Demonstrations


1️⃣ 一句话总结

这篇论文通过构建一个高质量的大规模桌面操作数据集GroundCUA,并训练出高效的GroundNext模型,显著提升了计算机使用智能体将语言指令准确对应到屏幕元素的能力,同时大幅减少了所需训练数据量。


📄 打开原文 PDF