菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-23
📄 Abstract - Goal-Oriented Influence-Maximizing Data Acquisition for Learning and Optimization

Active data acquisition is central to many learning and optimization tasks in deep neural networks, yet remains challenging because most approaches rely on predictive uncertainty estimates that are difficult to obtain reliably. To this end, we propose Goal-Oriented Influence- Maximizing Data Acquisition (GOIMDA), an active acquisition algorithm that avoids explicit posterior inference while remaining uncertainty-aware through inverse curvature. GOIMDA selects inputs by maximizing their expected influence on a user-specified goal functional, such as test loss, predictive entropy, or the value of an optimizer-recommended design. Leveraging first-order influence functions, we derive a tractable acquisition rule that combines the goal gradient, training-loss curvature, and candidate sensitivity to model parameters. We show theoretically that, for generalized linear models, GOIMDA approximates predictive-entropy minimization up to a correction term accounting for goal alignment and prediction bias, thereby, yielding uncertainty-aware behavior without maintaining a Bayesian posterior. Empirically, across learning tasks (including image and text classification) and optimization tasks (including noisy global optimization benchmarks and neural-network hyperparameter tuning), GOIMDA consistently reaches target performance with substantially fewer labeled samples or function evaluations than uncertainty-based active learning and Gaussian-process Bayesian optimization baselines.

顶级标签: machine learning model training data
详细标签: active learning data acquisition influence functions optimization curvature 或 搜索:

面向目标的影响力最大化数据采集用于学习与优化 / Goal-Oriented Influence-Maximizing Data Acquisition for Learning and Optimization


1️⃣ 一句话总结

这篇论文提出了一种名为GOIMDA的新型主动数据采集算法,它通过最大化所选数据对用户指定目标(如测试误差或优化器推荐值)的预期影响力来高效选择数据,无需复杂的概率推断,就能在多种学习和优化任务中用更少的样本达到目标性能。

源自 arXiv: 2602.19578