📄
Abstract - To Move or Not to Move: Constraint-based Planning Enables Zero-Shot Generalization for Interactive Navigation
Visual navigation typically assumes the existence of at least one obstacle-free path between start and goal, which must be discovered/planned by the robot. However, in real-world scenarios, such as home environments and warehouses, clutter can block all routes. Targeted at such cases, we introduce the Lifelong Interactive Navigation problem, where a mobile robot with manipulation abilities can move clutter to forge its own path to complete sequential object- placement tasks - each involving placing an given object (eg. Alarm clock, Pillow) onto a target object (eg. Dining table, Desk, Bed). To address this lifelong setting - where effects of environment changes accumulate and have long-term effects - we propose an LLM-driven, constraint-based planning framework with active perception. Our framework allows the LLM to reason over a structured scene graph of discovered objects and obstacles, deciding which object to move, where to place it, and where to look next to discover task-relevant information. This coupling of reasoning and active perception allows the agent to explore the regions expected to contribute to task completion rather than exhaustively mapping the environment. A standard motion planner then executes the corresponding navigate-pick-place, or detour sequence, ensuring reliable low-level control. Evaluated in physics-enabled ProcTHOR-10k simulator, our approach outperforms non-learning and learning-based baselines. We further demonstrate our approach qualitatively on real-world hardware.
移动还是不移动:基于约束的规划实现交互式导航的零样本泛化 /
To Move or Not to Move: Constraint-based Planning Enables Zero-Shot Generalization for Interactive Navigation
1️⃣ 一句话总结
这篇论文提出了一种让机器人能够主动移动障碍物来开辟新路径的智能导航方法,通过结合大语言模型的推理和主动感知能力,使机器人能在复杂、杂乱的环境中完成一系列物品摆放任务,并在模拟和真实硬件中验证了其有效性。