菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-14
📄 Abstract - CLARE: Continual Learning for Vision-Language-Action Models via Autonomous Adapter Routing and Expansion

To teach robots complex manipulation tasks, it is now a common practice to fine-tune a pre-trained vision-language-action model (VLA) on task-specific data. However, since this recipe updates existing representations, it is unsuitable for long-term operation in the real world, where robots must continually adapt to new tasks and environments while retaining the knowledge they have already acquired. Existing continual learning methods for robotics commonly require storing previous data (exemplars), struggle with long task sequences, or rely on task identifiers for deployment. To address these limitations, we propose CLARE, a general, parameter-efficient framework for exemplar-free continual learning with VLAs. CLARE introduces lightweight modular adapters into selected feedforward layers and autonomously expands the model only where necessary when learning a new task, guided by layer-wise feature similarity. During deployment, an autoencoder-based routing mechanism dynamically activates the most relevant adapters without requiring task labels. Through extensive experiments on the LIBERO benchmark, we show that CLARE achieves high performance on new tasks without catastrophic forgetting of earlier tasks, significantly outperforming even exemplar-based methods. Code and data are available at this https URL.

顶级标签: robotics multi-modal model training
详细标签: continual learning vision-language-action adapter routing parameter-efficient catastrophic forgetting 或 搜索:

CLARE:通过自主适配器路由与扩展实现视觉-语言-动作模型的持续学习 / CLARE: Continual Learning for Vision-Language-Action Models via Autonomous Adapter Routing and Expansion


1️⃣ 一句话总结

这篇论文提出了一种名为CLARE的新方法,让机器人能够在不断学习新技能的同时牢牢记住旧本领,而且不需要存储旧数据或依赖任务标签,从而更灵活、高效地适应真实世界的长期运行。

源自 arXiv: 2601.09512