何时行动、询问或学习:基于不确定性的策略引导 / When to Act, Ask, or Learn: Uncertainty-Aware Policy Steering
1️⃣ 一句话总结
这篇论文提出了一个名为UPS的智能机器人决策框架,它能够识别任务指令的模糊性和自身执行能力的不确定性,从而自主选择是执行动作、向人类询问澄清,还是请求干预学习,以此减少对人工反馈的依赖并提升执行可靠性。
Policy steering is an emerging way to adapt robot behaviors at deployment-time: a learned verifier analyzes low-level action samples proposed by a pre-trained policy (e.g., diffusion policy) and selects only those aligned with the task. While Vision-Language Models (VLMs) are promising general-purpose verifiers due to their reasoning capabilities, existing frameworks often assume these models are well-calibrated. In practice, the overconfident judgment from VLM can degrade the steering performance under both high-level semantic uncertainty in task specifications and low-level action uncertainty or incapability of the pre-trained policy. We propose uncertainty-aware policy steering (UPS), a framework that jointly reasons about semantic task uncertainty and low-level action feasibility, and selects an uncertainty resolution strategy: execute a high-confidence action, clarify task ambiguity via natural language queries, or ask for action interventions to correct the low-level policy when it is deemed incapable at the task. We leverage conformal prediction to calibrate the composition of the VLM and the pre-trained base policy, providing statistical assurances that the verifier selects the correct strategy. After collecting interventions during deployment, we employ residual learning to improve the capability of the pre-trained policy, enabling the system to learn continually but with minimal expensive human feedback. We demonstrate our framework through experiments in simulation and on hardware, showing that UPS can disentangle confident, ambiguous, and incapable scenarios and minimizes expensive user interventions compared to uncalibrated baselines and prior human- or robot-gated continual learning approaches. Videos can be found at this https URL
何时行动、询问或学习:基于不确定性的策略引导 / When to Act, Ask, or Learn: Uncertainty-Aware Policy Steering
这篇论文提出了一个名为UPS的智能机器人决策框架,它能够识别任务指令的模糊性和自身执行能力的不确定性,从而自主选择是执行动作、向人类询问澄清,还是请求干预学习,以此减少对人工反馈的依赖并提升执行可靠性。
源自 arXiv: 2602.22474