菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-30
📄 Abstract - GR-Dexter Technical Report

Vision-language-action (VLA) models have enabled language-conditioned, long-horizon robot manipulation, but most existing systems are limited to grippers. Scaling VLA policies to bimanual robots with high degree-of-freedom (DoF) dexterous hands remains challenging due to the expanded action space, frequent hand-object occlusions, and the cost of collecting real-robot data. We present GR-Dexter, a holistic hardware-model-data framework for VLA-based generalist manipulation on a bimanual dexterous-hand robot. Our approach combines the design of a compact 21-DoF robotic hand, an intuitive bimanual teleoperation system for real-robot data collection, and a training recipe that leverages teleoperated robot trajectories together with large-scale vision-language and carefully curated cross-embodiment datasets. Across real-world evaluations spanning long-horizon everyday manipulation and generalizable pick-and-place, GR-Dexter achieves strong in-domain performance and improved robustness to unseen objects and unseen instructions. We hope GR-Dexter serves as a practical step toward generalist dexterous-hand robotic manipulation.

顶级标签: robotics multi-modal agents
详细标签: vision-language-action dexterous manipulation bimanual robot teleoperation real-robot data 或 搜索:

GR-Dexter技术报告 / GR-Dexter Technical Report


1️⃣ 一句话总结

这篇论文提出了一个名为GR-Dexter的软硬件一体化框架,它通过设计灵巧的双手机器人、便捷的遥操作系统以及创新的数据训练方法,成功实现了让机器人能像人一样根据语言指令完成各种复杂的双手操作任务。

源自 arXiv: 2512.24210