菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-17
📄 Abstract - Contextual Preference Distribution Learning

Decision-making problems often feature uncertainty stemming from heterogeneous and context-dependent human preferences. To address this, we propose a sequential learning-and-optimization pipeline to learn preference distributions and leverage them to solve downstream problems, for example risk-averse formulations. We focus on human choice settings that can be formulated as (integer) linear programs. In such settings, existing inverse optimization and choice modelling methods infer preferences from observed choices but typically produce point estimates or fail to capture contextual shifts, making them unsuitable for risk-averse decision-making. Using a bounded-variance score function gradient estimator, we train a predictive model mapping contextual features to a rich class of parameterizable distributions. This approach yields a maximum likelihood estimate. The model generates scenarios for unseen contexts in the subsequent optimization phase. In a synthetic ridesharing environment, our approach reduces average post-decision surprise by up to 114$\times$ compared to a risk-neutral approach with perfect predictions and up to 25$\times$ compared to leading risk-averse baselines.

顶级标签: machine learning systems theory
详细标签: preference learning inverse optimization risk-averse decision-making contextual modeling sequential optimization 或 搜索:

情境偏好分布学习 / Contextual Preference Distribution Learning


1️⃣ 一句话总结

这篇论文提出了一种新的学习框架,通过捕捉人类决策中复杂多变且受情境影响的偏好分布,并将其用于后续的风险规避优化问题,从而显著降低了决策后的意外风险。

源自 arXiv: 2603.17139