菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-03
📄 Abstract - Directional Neural Collapse Explains Few-Shot Transfer in Self-Supervised Learning

Frozen self-supervised representations often transfer well with only a few labels across many semantic tasks. We argue that a single geometric quantity, \emph{directional} CDNV (decision-axis variance), sits at the core of two favorable behaviors: strong few-shot transfer within a task, and low interference across many tasks. We show that both emerge when variability \emph{along} class-separating directions is small. First, we prove sharp non-asymptotic multiclass generalization bounds for downstream classification whose leading term is the directional CDNV. The bounds include finite-shot corrections that cleanly separate intrinsic decision-axis variability from centroid-estimation error. Second, we link decision-axis collapse to multitask geometry: for independent balanced labelings, small directional CDNV across tasks forces the corresponding decision axes to be nearly orthogonal, helping a single representation support many tasks with minimal interference. Empirically, across SSL objectives, directional CDNV collapses during pretraining even when classical CDNV remains large, and our bounds closely track few-shot error at practical shot sizes. Additionally, on synthetic multitask data, we verify that SSL learns representations whose induced decision axes are nearly orthogonal. The code and project page of the paper are available at [\href{this https URL}{project page}].

顶级标签: machine learning model evaluation theory
详细标签: self-supervised learning few-shot transfer neural collapse representation geometry generalization bounds 或 搜索:

方向性神经坍缩解释自监督学习中的小样本迁移 / Directional Neural Collapse Explains Few-Shot Transfer in Self-Supervised Learning


1️⃣ 一句话总结

这篇论文发现,自监督学习模型之所以能用很少的标签就在多个任务上表现良好,关键在于其内部特征沿着分类决策方向的变化很小,这既保证了单个任务的小样本学习能力,也使得不同任务的决策方向几乎正交,从而减少了任务间的干扰。

源自 arXiv: 2603.03530