菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-28
📄 Abstract - Active In-Context Learning for Tabular Foundation Models

Active learning (AL) reduces labeling cost by querying informative samples, but in tabular settings its cold-start gains are often limited because uncertainty estimates are unreliable when models are trained on very few labels. Tabular foundation models such as TabPFN provide calibrated probabilistic predictions via in-context learning (ICL), i.e., without task-specific weight updates, enabling an AL regime in which the labeled context - rather than parameters - is iteratively optimized. We formalize Tabular Active In-Context Learning (Tab-AICL) and instantiate it with four acquisition rules: uncertainty (TabPFN-Margin), diversity (TabPFN-Coreset), an uncertainty-diversity hybrid (TabPFN-Hybrid), and a scalable two-stage method (TabPFN-Proxy-Hybrid) that shortlists candidates using a lightweight linear proxy before TabPFN-based selection. Across 20 classification benchmarks, Tab-AICL improves cold-start sample efficiency over retrained gradient-boosting baselines (CatBoost-Margin and XGBoost-Margin), measured by normalized AULC up to 100 labeled samples.

顶级标签: machine learning model training data
详细标签: active learning tabular data in-context learning foundation models sample efficiency 或 搜索:

面向表格基础模型的主动上下文学习 / Active In-Context Learning for Tabular Foundation Models


1️⃣ 一句话总结

这篇论文提出了一种名为Tab-AICL的新方法,它结合了主动学习和上下文学习,让表格基础模型(如TabPFN)在只有少量标注数据时,也能高效地选择最有价值的样本进行标注,从而在冷启动阶段显著提升了学习效率。

源自 arXiv: 2603.27385