RDT2:探索UMI数据的规模极限以实现零样本跨硬件平台泛化 / RDT2: Exploring the Scaling Limit of UMI Data Towards Zero-Shot Cross-Embodiment Generalization
1️⃣ 一句话总结
这篇论文提出了一个名为RDT2的机器人基础模型,它通过收集大规模通用数据集和创新的训练方法,首次实现了无需额外训练就能让机器人完成新任务、适应新场景、甚至操控从未见过的硬件平台。
Vision-Language-Action (VLA) models hold promise for generalist robotics but currently struggle with data scarcity, architectural inefficiencies, and the inability to generalize across different hardware platforms. We introduce RDT2, a robotic foundation model built upon a 7B parameter VLM designed to enable zero-shot deployment on novel embodiments for open-vocabulary tasks. To achieve this, we collected one of the largest open-source robotic datasets--over 10,000 hours of demonstrations in diverse families--using an enhanced, embodiment-agnostic Universal Manipulation Interface (UMI). Our approach employs a novel three-stage training recipe that aligns discrete linguistic knowledge with continuous control via Residual Vector Quantization (RVQ), flow-matching, and distillation for real-time inference. Consequently, RDT2 becomes one of the first models that simultaneously zero-shot generalizes to unseen objects, scenes, instructions, and even robotic platforms. Besides, it outperforms state-of-the-art baselines in dexterous, long-horizon, and dynamic downstream tasks like playing table tennis. See this https URL for more information.
RDT2:探索UMI数据的规模极限以实现零样本跨硬件平台泛化 / RDT2: Exploring the Scaling Limit of UMI Data Towards Zero-Shot Cross-Embodiment Generalization
这篇论文提出了一个名为RDT2的机器人基础模型,它通过收集大规模通用数据集和创新的训练方法,首次实现了无需额外训练就能让机器人完成新任务、适应新场景、甚至操控从未见过的硬件平台。
源自 arXiv: 2602.03310