DualCoT-VLA:通过并行推理实现视觉-语言-动作模型的视觉语言思维链 / DualCoT-VLA: Visual-Linguistic Chain of Thought via Parallel Reasoning for Vision-Language-Action Models
1️⃣ 一句话总结
这篇论文提出了一种名为DualCoT-VLA的新方法,它通过并行的视觉和语言思维链,让机器人模型能够同时进行精细的空间感知和高级任务规划,从而更快速、更准确地完成复杂的多步骤操作任务。
Vision-Language-Action (VLA) models map visual observations and language instructions directly to robotic actions. While effective for simple tasks, standard VLA models often struggle with complex, multi-step tasks requiring logical planning, as well as precise manipulations demanding fine-grained spatial perception. Recent efforts have incorporated Chain-of-Thought (CoT) reasoning to endow VLA models with a ``thinking before acting'' capability. However, current CoT-based VLA models face two critical limitations: 1) an inability to simultaneously capture low-level visual details and high-level logical planning due to their reliance on isolated, single-modal CoT; 2) high inference latency with compounding errors caused by step-by-step autoregressive decoding. To address these limitations, we propose DualCoT-VLA, a visual-linguistic CoT method for VLA models with a parallel reasoning mechanism. To achieve comprehensive multi-modal reasoning, our method integrates a visual CoT for low-level spatial understanding and a linguistic CoT for high-level task planning. Furthermore, to overcome the latency bottleneck, we introduce a parallel CoT mechanism that incorporates two sets of learnable query tokens, shifting autoregressive reasoning to single-step forward reasoning. Extensive experiments demonstrate that our DualCoT-VLA achieves state-of-the-art performance on the LIBERO and RoboCasa GR1 benchmarks, as well as in real-world platforms.
DualCoT-VLA:通过并行推理实现视觉-语言-动作模型的视觉语言思维链 / DualCoT-VLA: Visual-Linguistic Chain of Thought via Parallel Reasoning for Vision-Language-Action Models
这篇论文提出了一种名为DualCoT-VLA的新方法,它通过并行的视觉和语言思维链,让机器人模型能够同时进行精细的空间感知和高级任务规划,从而更快速、更准确地完成复杂的多步骤操作任务。
源自 arXiv: 2603.22280