菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-14
📄 Abstract - Advancing Analytic Class-Incremental Learning through Vision-Language Calibration

Class-incremental learning (CIL) with pre-trained models (PTMs) faces a critical trade-off between efficient adaptation and long-term stability. While analytic learning enables rapid, recursive closed-form updates, its efficacy is often compromised by accumulated errors and feature incompatibility. In this paper, we first conduct a systematic study to dissect the failure modes of PTM-based analytic CIL, identifying representation rigidity as the primary bottleneck. Motivated by these insights, we propose \textbf{VILA}, a novel dual-branch framework that advances analytic CIL via a two-level vision-language calibration strategy. Specifically, we coherently fuse plastic, task-adapted features with a frozen, universal semantic anchor at the feature level through geometric calibration, and leverage cross-modal priors at the decision level to rectify prediction bias. This confluence maintains analytic-learning's extreme efficiency while overcoming its inherent brittleness. Extensive experiments across eight benchmarks demonstrate that VILA consistently yields superior performance, particularly in fine-grained and long-sequence scenarios. Our framework harmonizes high-fidelity prediction with the simplicity of analytic learning. Our code is available at this https URL

顶级标签: model training computer vision multi-modal
详细标签: class-incremental learning vision-language models analytic learning representation calibration fine-grained classification 或 搜索:

通过视觉-语言校准推进解析式类增量学习 / Advancing Analytic Class-Incremental Learning through Vision-Language Calibration


1️⃣ 一句话总结

这篇论文提出了一种名为VILA的新方法,它巧妙地结合了视觉和语言信息来校准模型,从而在让AI系统高效学习新类别的同时,有效防止了遗忘旧知识并提升了长期稳定性。

源自 arXiv: 2602.13670