菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-03
📄 Abstract - Tether: Autonomous Functional Play with Correspondence-Driven Trajectory Warping

The ability to conduct and learn from interaction and experience is a central challenge in robotics, offering a scalable alternative to labor-intensive human demonstrations. However, realizing such &#34;play&#34; requires (1) a policy robust to diverse, potentially out-of-distribution environment states, and (2) a procedure that continuously produces useful robot experience. To address these challenges, we introduce Tether, a method for autonomous functional play involving structured, task-directed interactions. First, we design a novel open-loop policy that warps actions from a small set of source demonstrations (<=10) by anchoring them to semantic keypoint correspondences in the target scene. We show that this design is extremely data-efficient and robust even under significant spatial and semantic variations. Second, we deploy this policy for autonomous functional play in the real world via a continuous cycle of task selection, execution, evaluation, and improvement, guided by the visual understanding capabilities of vision-language models. This procedure generates diverse, high-quality datasets with minimal human intervention. In a household-like multi-object setup, our method is the first to perform many hours of autonomous multi-task play in the real world starting from only a handful of demonstrations. This produces a stream of data that consistently improves the performance of closed-loop imitation policies over time, ultimately yielding over 1000 expert-level trajectories and training policies competitive with those learned from human-collected demonstrations.

顶级标签: robotics agents model training
详细标签: autonomous play trajectory warping keypoint correspondences imitation learning vision-language models 或 搜索:

Tether:基于对应关系驱动轨迹扭曲的自主功能化交互学习 / Tether: Autonomous Functional Play with Correspondence-Driven Trajectory Warping


1️⃣ 一句话总结

这篇论文提出了一种名为Tether的新方法,它能让机器人像玩游戏一样自主探索和学习新任务:只需少量演示,机器人就能通过视觉关键点匹配来调整动作,并在现实环境中持续执行、评估和改进任务,从而自动生成大量高质量的训练数据来提升自身技能。

源自 arXiv: 2603.03278