EVOLVE-VLA:基于环境反馈的视觉-语言-动作模型测试时训练框架 / EVOLVE-VLA: Test-Time Training from Environment Feedback for Vision-Language-Action Models
1️⃣ 一句话总结
这篇论文提出了一个名为EVOLVE-VLA的新框架,让机器人视觉-语言-动作模型能在实际测试环境中通过与环境的互动来自主学习和改进,从而显著减少对大量人工演示数据的依赖,并提升其在陌生任务和复杂长序列任务中的适应能力。
Achieving truly adaptive embodied intelligence requires agents that learn not just by imitating static demonstrations, but by continuously improving through environmental interaction, which is akin to how humans master skills through practice. Vision-Language-Action (VLA) models have advanced robotic manipulation by leveraging large language models, yet remain fundamentally limited by Supervised Finetuning (SFT): requiring hundreds of demonstrations per task, rigidly memorizing trajectories, and failing to adapt when deployment conditions deviate from training. We introduce EVOLVE-VLA, a test-time training framework enabling VLAs to continuously adapt through environment interaction with minimal or zero task-specific demonstrations. The key technical challenge is replacing oracle reward signals (unavailable at test time) with autonomous feedback. We address this through a learned progress estimator providing dense feedback, and critically, we design our framework to ``tame'' this inherently noisy signal via two mechanisms: (1) an accumulative progress estimation mechanism smoothing noisy point-wise estimates, and (2) a progressive horizon extension strategy enabling gradual policy evolution. EVOLVE-VLA achieves substantial gains: +8.6\% on long-horizon tasks, +22.0\% in 1-shot learning, and enables cross-task generalization -- achieving 20.8\% success on unseen tasks without task-specific demonstrations training (vs. 0\% for pure SFT). Qualitative analysis reveals emergent capabilities absent in demonstrations, including error recovery and novel strategies. This work represents a critical step toward VLAs that truly learn and adapt, moving beyond static imitation toward continuous self-improvements.
EVOLVE-VLA:基于环境反馈的视觉-语言-动作模型测试时训练框架 / EVOLVE-VLA: Test-Time Training from Environment Feedback for Vision-Language-Action Models
这篇论文提出了一个名为EVOLVE-VLA的新框架,让机器人视觉-语言-动作模型能在实际测试环境中通过与环境的互动来自主学习和改进,从而显著减少对大量人工演示数据的依赖,并提升其在陌生任务和复杂长序列任务中的适应能力。
源自 arXiv: 2512.14666